var/home/core/zuul-output/0000755000175000017500000000000015113247154014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113260652015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005312725515113260644017711 0ustar rootrootDec 01 08:17:47 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 08:17:47 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:17:47 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 08:17:48 crc kubenswrapper[4744]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:17:48 crc kubenswrapper[4744]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 08:17:48 crc kubenswrapper[4744]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:17:48 crc kubenswrapper[4744]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:17:48 crc kubenswrapper[4744]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 08:17:48 crc kubenswrapper[4744]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.138388 4744 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141141 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141158 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141164 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141169 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141173 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141178 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141183 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141188 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141192 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141197 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141201 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141205 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141210 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141214 4744 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141231 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141236 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141240 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141245 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141249 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141253 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141258 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141262 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141267 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141271 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141276 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141281 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141285 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141290 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141295 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141301 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141306 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141311 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141318 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141324 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141330 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141335 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141341 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141345 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141350 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141356 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141360 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141365 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141370 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141375 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141380 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141385 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141390 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141394 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141399 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141419 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141436 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141441 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141446 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141450 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141456 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141462 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141468 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141472 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141477 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141484 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141488 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141494 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141500 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141504 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141510 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141514 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141519 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141524 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141529 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141534 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.141539 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141637 4744 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141648 4744 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141661 4744 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141669 4744 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141678 4744 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141683 4744 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141691 4744 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141697 4744 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141702 4744 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141707 4744 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141714 4744 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141719 4744 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141724 4744 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141729 4744 flags.go:64] FLAG: --cgroup-root="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141735 4744 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141748 4744 flags.go:64] FLAG: --client-ca-file="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141753 4744 flags.go:64] FLAG: --cloud-config="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141758 4744 flags.go:64] FLAG: --cloud-provider="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141763 4744 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141773 4744 flags.go:64] FLAG: --cluster-domain="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141778 4744 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141783 4744 flags.go:64] FLAG: --config-dir="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141789 4744 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141794 4744 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141802 4744 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141807 4744 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141813 4744 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141818 4744 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141824 4744 flags.go:64] FLAG: --contention-profiling="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141829 4744 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141834 4744 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141841 4744 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141846 4744 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141852 4744 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141858 4744 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141863 4744 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141868 4744 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141873 4744 flags.go:64] FLAG: --enable-server="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141878 4744 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141887 4744 flags.go:64] FLAG: --event-burst="100" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141893 4744 flags.go:64] FLAG: --event-qps="50" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141898 4744 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141903 4744 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141909 4744 flags.go:64] FLAG: --eviction-hard="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141915 4744 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141921 4744 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141926 4744 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141931 4744 flags.go:64] FLAG: --eviction-soft="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141936 4744 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141941 4744 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141946 4744 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141954 4744 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141960 4744 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141965 4744 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141970 4744 flags.go:64] FLAG: --feature-gates="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141977 4744 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141982 4744 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141988 4744 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141993 4744 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.141998 4744 flags.go:64] FLAG: --healthz-port="10248" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142003 4744 flags.go:64] FLAG: --help="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142009 4744 flags.go:64] FLAG: --hostname-override="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142014 4744 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142020 4744 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142026 4744 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142031 4744 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142036 4744 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142042 4744 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142047 4744 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142052 4744 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142058 4744 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142064 4744 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142070 4744 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142075 4744 flags.go:64] FLAG: --kube-reserved="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142080 4744 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142085 4744 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142091 4744 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142097 4744 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142102 4744 flags.go:64] FLAG: --lock-file="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142107 4744 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142113 4744 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142119 4744 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142127 4744 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142134 4744 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142139 4744 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142145 4744 flags.go:64] FLAG: --logging-format="text" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142150 4744 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142157 4744 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142162 4744 flags.go:64] FLAG: --manifest-url="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142167 4744 flags.go:64] FLAG: --manifest-url-header="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142174 4744 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142179 4744 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142186 4744 flags.go:64] FLAG: --max-pods="110" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142191 4744 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142196 4744 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142202 4744 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142207 4744 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142213 4744 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142218 4744 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142223 4744 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142236 4744 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142241 4744 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142246 4744 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142252 4744 flags.go:64] FLAG: --pod-cidr="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142272 4744 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142282 4744 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142287 4744 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142293 4744 flags.go:64] FLAG: --pods-per-core="0" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142298 4744 flags.go:64] FLAG: --port="10250" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142304 4744 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142309 4744 flags.go:64] FLAG: --provider-id="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142314 4744 flags.go:64] FLAG: --qos-reserved="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142320 4744 flags.go:64] FLAG: --read-only-port="10255" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142325 4744 flags.go:64] FLAG: --register-node="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142330 4744 flags.go:64] FLAG: --register-schedulable="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142336 4744 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142364 4744 flags.go:64] FLAG: --registry-burst="10" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142369 4744 flags.go:64] FLAG: --registry-qps="5" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142375 4744 flags.go:64] FLAG: --reserved-cpus="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142380 4744 flags.go:64] FLAG: --reserved-memory="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142386 4744 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142392 4744 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142397 4744 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142438 4744 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142444 4744 flags.go:64] FLAG: --runonce="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142450 4744 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142455 4744 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142461 4744 flags.go:64] FLAG: --seccomp-default="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142467 4744 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142472 4744 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142477 4744 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142482 4744 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142488 4744 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142493 4744 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142498 4744 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142503 4744 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142508 4744 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142514 4744 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142519 4744 flags.go:64] FLAG: --system-cgroups="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142524 4744 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142533 4744 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142538 4744 flags.go:64] FLAG: --tls-cert-file="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142543 4744 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142550 4744 flags.go:64] FLAG: --tls-min-version="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142555 4744 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142560 4744 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142565 4744 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142571 4744 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142576 4744 flags.go:64] FLAG: --v="2" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142583 4744 flags.go:64] FLAG: --version="false" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142590 4744 flags.go:64] FLAG: --vmodule="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142596 4744 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.142602 4744 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142735 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142744 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142750 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142754 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142759 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142764 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142770 4744 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142778 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142783 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142787 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142792 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142797 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142802 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142806 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142811 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142816 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142820 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142825 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142829 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142834 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142838 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142842 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142849 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142855 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142859 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142865 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142869 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142874 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142878 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142884 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142890 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142895 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142900 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142905 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142909 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142914 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142918 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142923 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142928 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142935 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142940 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142946 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142952 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142957 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142962 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142968 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142972 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142977 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142981 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142986 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142990 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142994 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.142999 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143003 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143008 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143012 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143017 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143021 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143026 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143030 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143035 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143039 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143043 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143048 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143053 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143057 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143062 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143066 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143071 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143077 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.143082 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.143099 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.150309 4744 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.150359 4744 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150480 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150492 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150498 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150503 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150508 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150513 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150517 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150522 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150528 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150534 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150539 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150544 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150548 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150554 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150558 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150563 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150567 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150573 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150577 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150582 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150587 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150592 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150597 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150601 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150606 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150611 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150615 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150619 4744 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150624 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150629 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150633 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150638 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150642 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150657 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150662 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150667 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150671 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150675 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150680 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150685 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150689 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150693 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150698 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150702 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150707 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150712 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150716 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150720 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150725 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150730 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150736 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150742 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150748 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150753 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150758 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150763 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150768 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150773 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150777 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150781 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150787 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150791 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150796 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150800 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150804 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150809 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150813 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150818 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150822 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150835 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.150839 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.150847 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151002 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151011 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151015 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151021 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151027 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151033 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151038 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151043 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151048 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151052 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151058 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151063 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151068 4744 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151073 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151077 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151084 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151089 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151094 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151100 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151106 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151112 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151117 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151122 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151126 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151131 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151135 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151140 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151144 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151148 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151153 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151157 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151161 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151166 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151182 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151187 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151191 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151195 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151200 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151204 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151209 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151213 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151217 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151222 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151226 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151231 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151235 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151240 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151244 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151248 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151253 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151257 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151262 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151266 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151270 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151275 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151279 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151284 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151288 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151293 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151297 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151302 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151306 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151312 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151317 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151322 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151326 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151331 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151335 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151340 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151352 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.151358 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.151366 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.151738 4744 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.157996 4744 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.158106 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.158841 4744 server.go:997] "Starting client certificate rotation" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.158873 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.159063 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-08 09:12:31.840349164 +0000 UTC Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.159189 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.165126 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.166395 4744 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.166803 4744 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.174500 4744 log.go:25] "Validated CRI v1 runtime API" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.195438 4744 log.go:25] "Validated CRI v1 image API" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.197072 4744 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.199586 4744 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-08-13-34-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.199622 4744 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.216142 4744 manager.go:217] Machine: {Timestamp:2025-12-01 08:17:48.214671741 +0000 UTC m=+0.203729682 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:734ce460-7574-4e1f-8e70-e2470c8ff4bb BootID:eae1d0fe-cf0c-4849-8221-3d8cb97835e0 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:23:7b:95 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:23:7b:95 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:da:2a:5e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:62:7c:41 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:6f:ab:f9 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ce:46:75 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ea:9e:47:4d:76:28 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:52:3d:43:43:f5:38 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.216395 4744 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.216637 4744 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.217326 4744 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.217536 4744 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.217574 4744 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.219756 4744 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.219884 4744 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.220221 4744 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.220448 4744 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.220751 4744 state_mem.go:36] "Initialized new in-memory state store" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.220921 4744 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.221693 4744 kubelet.go:418] "Attempting to sync node with API server" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.221802 4744 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.221922 4744 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.222011 4744 kubelet.go:324] "Adding apiserver pod source" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.222158 4744 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.223542 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.223623 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.223653 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.223755 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.224401 4744 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.224778 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.225593 4744 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226234 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226264 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226277 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226288 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226307 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226318 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226330 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226347 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226362 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226373 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226388 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226400 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.226874 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.227318 4744 server.go:1280] "Started kubelet" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.227750 4744 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.227775 4744 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.228214 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:48 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230021 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230050 4744 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230183 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 17:20:20.84131603 +0000 UTC Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230209 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 633h2m32.611108719s for next certificate rotation Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230291 4744 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230311 4744 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.230364 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230438 4744 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.230720 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.230775 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.230966 4744 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.231223 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="200ms" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.231429 4744 server.go:460] "Adding debug handlers to kubelet server" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.233235 4744 factory.go:55] Registering systemd factory Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.233273 4744 factory.go:221] Registration of the systemd container factory successfully Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.233657 4744 factory.go:153] Registering CRI-O factory Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.232613 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d097e927cdffe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:17:48.227289086 +0000 UTC m=+0.216347017,LastTimestamp:2025-12-01 08:17:48.227289086 +0000 UTC m=+0.216347017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.233681 4744 factory.go:221] Registration of the crio container factory successfully Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.233804 4744 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.233845 4744 factory.go:103] Registering Raw factory Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.233875 4744 manager.go:1196] Started watching for new ooms in manager Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.235209 4744 manager.go:319] Starting recovery of all containers Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244611 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244695 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244718 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244736 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244756 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244775 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244794 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244811 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244834 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244851 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244870 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244887 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244905 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244975 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.244995 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245013 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245030 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245050 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245068 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245087 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245104 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245121 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245139 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245157 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245175 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245192 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245214 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245235 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245252 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245268 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245308 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245326 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245345 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245362 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245396 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245449 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245473 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245491 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245511 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245529 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245547 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245566 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245584 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245604 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245621 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245640 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245661 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245677 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245695 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245712 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245730 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245748 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245771 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245791 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245811 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245831 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245849 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245868 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245886 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245903 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245946 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245966 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.245984 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246000 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246017 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246061 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246079 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246098 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246115 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246135 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246158 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246182 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246206 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246223 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246240 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246258 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246277 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246295 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246314 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246333 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246351 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246368 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246386 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246436 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246465 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246484 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246502 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246520 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246538 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246556 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246573 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246593 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246611 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246629 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246650 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246669 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246686 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246707 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246733 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246758 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246782 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246806 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246832 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246857 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246890 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246917 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246945 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246973 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.246999 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.247025 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.247056 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.247081 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.247106 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.247131 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.247161 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248093 4744 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248143 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248173 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248201 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248227 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248252 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248278 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248306 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248364 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248390 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248452 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248485 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248511 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248535 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248559 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248581 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248605 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248628 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248652 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248675 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248700 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248723 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248759 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248784 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248810 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248834 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248860 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248886 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248910 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248934 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248957 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.248982 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249007 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249033 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249057 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249081 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249115 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249141 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249170 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249194 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249217 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249239 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249264 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249287 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249310 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249334 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249356 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249379 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249401 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249462 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249487 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249511 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249534 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249557 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249580 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249604 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249625 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249649 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249672 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249697 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249721 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249743 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249768 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249792 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249815 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249838 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249860 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249882 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249904 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249931 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249974 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.249998 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250044 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250068 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250091 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250127 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250150 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250172 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250195 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250218 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250241 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250263 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250287 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250311 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250334 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250357 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250381 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250430 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250457 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250481 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250504 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250532 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250556 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250578 4744 reconstruct.go:97] "Volume reconstruction finished" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.250593 4744 reconciler.go:26] "Reconciler: start to sync state" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.262049 4744 manager.go:324] Recovery completed Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.275530 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.279555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.279611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.279620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.280959 4744 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.280982 4744 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.281005 4744 state_mem.go:36] "Initialized new in-memory state store" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.281315 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.283691 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.283732 4744 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.283756 4744 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.283795 4744 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.284725 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.284803 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.288777 4744 policy_none.go:49] "None policy: Start" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.289339 4744 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.289365 4744 state_mem.go:35] "Initializing new in-memory state store" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.330521 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.356792 4744 manager.go:334] "Starting Device Plugin manager" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.356839 4744 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.356853 4744 server.go:79] "Starting device plugin registration server" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.357248 4744 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.357263 4744 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.357469 4744 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.357541 4744 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.357548 4744 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.367247 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.384638 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.384754 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.386269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.386305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.386315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.386431 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.386865 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.386937 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.387381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.387452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.387470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.387629 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.387721 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.387751 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.388220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.388563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.388598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.391256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.391298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.391317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.391925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.392039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.392063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.392938 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.393194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.393267 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394477 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394486 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394504 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394735 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.394801 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395629 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395683 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.395939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.396811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.396872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.396896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.431850 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="400ms" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.456677 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.456790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.456834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.456873 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.456913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.456954 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457032 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457087 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457123 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457153 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457182 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457238 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457271 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457306 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457339 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.457361 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.458666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.458723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.458741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.458822 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.459366 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559087 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559161 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559199 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559262 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559292 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559357 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559397 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559494 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559473 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559503 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559557 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559583 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559560 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559557 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559598 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559528 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559387 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559873 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559907 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559976 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.560015 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.559915 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.659554 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.661322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.661366 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.661382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.661448 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.662100 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.717172 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.741767 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.741887 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.752505 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8043de1ea142822e7ab3bd29409ac40a6d269494af1d7f5a3343ff4134ad80bb WatchSource:0}: Error finding container 8043de1ea142822e7ab3bd29409ac40a6d269494af1d7f5a3343ff4134ad80bb: Status 404 returned error can't find the container with id 8043de1ea142822e7ab3bd29409ac40a6d269494af1d7f5a3343ff4134ad80bb Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.761145 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: I1201 08:17:48.767142 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.771376 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-64ca5f8bbe68961a31209867f0a559d637815da9ac1b031be01e521dc457b620 WatchSource:0}: Error finding container 64ca5f8bbe68961a31209867f0a559d637815da9ac1b031be01e521dc457b620: Status 404 returned error can't find the container with id 64ca5f8bbe68961a31209867f0a559d637815da9ac1b031be01e521dc457b620 Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.780964 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f239271e7c3f26db8ceeed2a419aa1a9fc97245ca230024bd40cdedc0d93e845 WatchSource:0}: Error finding container f239271e7c3f26db8ceeed2a419aa1a9fc97245ca230024bd40cdedc0d93e845: Status 404 returned error can't find the container with id f239271e7c3f26db8ceeed2a419aa1a9fc97245ca230024bd40cdedc0d93e845 Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.781937 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-97a49522a3063fa59334071293a44b7a20f16101dbb7f525892c1d5f45a8d464 WatchSource:0}: Error finding container 97a49522a3063fa59334071293a44b7a20f16101dbb7f525892c1d5f45a8d464: Status 404 returned error can't find the container with id 97a49522a3063fa59334071293a44b7a20f16101dbb7f525892c1d5f45a8d464 Dec 01 08:17:48 crc kubenswrapper[4744]: W1201 08:17:48.789247 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-cb2da444cb04c8caaaeba74244ee9b046105a4617e39386d31a0a5304f793d61 WatchSource:0}: Error finding container cb2da444cb04c8caaaeba74244ee9b046105a4617e39386d31a0a5304f793d61: Status 404 returned error can't find the container with id cb2da444cb04c8caaaeba74244ee9b046105a4617e39386d31a0a5304f793d61 Dec 01 08:17:48 crc kubenswrapper[4744]: E1201 08:17:48.833452 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="800ms" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.062415 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.063765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.063800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.063812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.063836 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:17:49 crc kubenswrapper[4744]: E1201 08:17:49.064288 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Dec 01 08:17:49 crc kubenswrapper[4744]: W1201 08:17:49.214084 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:49 crc kubenswrapper[4744]: E1201 08:17:49.214183 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.229568 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.292788 4744 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7" exitCode=0 Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.292914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.293056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cb2da444cb04c8caaaeba74244ee9b046105a4617e39386d31a0a5304f793d61"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.293179 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.294710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.294760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.294780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.299131 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.299180 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"97a49522a3063fa59334071293a44b7a20f16101dbb7f525892c1d5f45a8d464"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.301249 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e52d273edf1b7384ea71d7252be10669e628bb8e150d32e79f97898133c72a3e" exitCode=0 Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.301312 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e52d273edf1b7384ea71d7252be10669e628bb8e150d32e79f97898133c72a3e"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.301358 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f239271e7c3f26db8ceeed2a419aa1a9fc97245ca230024bd40cdedc0d93e845"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.301466 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.302186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.302215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.302225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.303571 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc" exitCode=0 Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.303633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.303647 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"64ca5f8bbe68961a31209867f0a559d637815da9ac1b031be01e521dc457b620"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.303734 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.304390 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.304425 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.304433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.306254 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e" exitCode=0 Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.306290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.306314 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8043de1ea142822e7ab3bd29409ac40a6d269494af1d7f5a3343ff4134ad80bb"} Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.307285 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.307961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.308013 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.308032 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:49 crc kubenswrapper[4744]: W1201 08:17:49.567252 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:49 crc kubenswrapper[4744]: E1201 08:17:49.567437 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:49 crc kubenswrapper[4744]: W1201 08:17:49.627767 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:49 crc kubenswrapper[4744]: E1201 08:17:49.627878 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:49 crc kubenswrapper[4744]: E1201 08:17:49.634693 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="1.6s" Dec 01 08:17:49 crc kubenswrapper[4744]: W1201 08:17:49.727796 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.39:6443: connect: connection refused Dec 01 08:17:49 crc kubenswrapper[4744]: E1201 08:17:49.727883 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.39:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.864653 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.866129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.866174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.866187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:49 crc kubenswrapper[4744]: I1201 08:17:49.866215 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:17:49 crc kubenswrapper[4744]: E1201 08:17:49.866769 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.39:6443: connect: connection refused" node="crc" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.289787 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.311171 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.311516 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.311531 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.311623 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.312515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.312546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.312557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.318277 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.318449 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.318504 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.318521 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.318979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.319009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.319022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.320543 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5068ca067105692d5312f839b3a530b1bfb5f4908f41ede747b9b5026b38b2a8" exitCode=0 Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.320650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5068ca067105692d5312f839b3a530b1bfb5f4908f41ede747b9b5026b38b2a8"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.320758 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.321462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.321481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.321493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324042 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324397 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918"} Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:50 crc kubenswrapper[4744]: I1201 08:17:50.324903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.330201 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e4d6a48cdf12f626971728b6129cdfdcf7ac240fd2d57f1bbd0769150bf46187" exitCode=0 Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.330313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e4d6a48cdf12f626971728b6129cdfdcf7ac240fd2d57f1bbd0769150bf46187"} Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.330519 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.332345 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.332454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.332479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.337125 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902"} Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.337210 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.339588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.339650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.339674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.341213 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc"} Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.341285 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.341347 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.343629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.343675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.343695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.343733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.343770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.343786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.467442 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.469227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.469304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.469321 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.469461 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:17:51 crc kubenswrapper[4744]: I1201 08:17:51.981755 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.350474 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c75976377e25c4dad812bf4645d19067e30a105e6f7d441f22cbe71f1bc8c6ab"} Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.350566 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.350564 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7eac91565f28c77e5ce801eceae446eb198e48a68054016d5a45f9869fc57171"} Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.350705 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.350741 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.350773 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3264a09da8207850acee482e1cfc3a509b5b85763227abf3a41512235545913e"} Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.350801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cb17b2f336bc4bff25a29bc95b8d11d380eaa6dc9984588aae31871d61e8928f"} Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.352084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.352144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.352169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.352233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.352288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:52 crc kubenswrapper[4744]: I1201 08:17:52.352310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.361274 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.362336 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.363022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1a16c0b2005868f2e7421bbf46c59cfb85bb00f878acaf39036f32f7d4eb5bb4"} Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.363666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.363724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.363746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.364622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.364672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:53 crc kubenswrapper[4744]: I1201 08:17:53.364697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.037467 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.064741 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.363858 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.363927 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.365642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.365642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.365737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.365755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.365703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.365818 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.524093 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.569815 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.570012 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.571496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.571550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:54 crc kubenswrapper[4744]: I1201 08:17:54.571588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.366569 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.366633 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.368339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.368391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.368449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.368460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.368474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:55 crc kubenswrapper[4744]: I1201 08:17:55.368482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.534925 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.535304 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.537362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.537580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.537709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.962087 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.962324 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.963836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.963879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:56 crc kubenswrapper[4744]: I1201 08:17:56.963897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:57 crc kubenswrapper[4744]: I1201 08:17:57.079957 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:57 crc kubenswrapper[4744]: I1201 08:17:57.087237 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:17:57 crc kubenswrapper[4744]: I1201 08:17:57.372545 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:57 crc kubenswrapper[4744]: I1201 08:17:57.373729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:57 crc kubenswrapper[4744]: I1201 08:17:57.373808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:57 crc kubenswrapper[4744]: I1201 08:17:57.373834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:58 crc kubenswrapper[4744]: E1201 08:17:58.368098 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 08:17:58 crc kubenswrapper[4744]: I1201 08:17:58.374473 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:17:58 crc kubenswrapper[4744]: I1201 08:17:58.375344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:17:58 crc kubenswrapper[4744]: I1201 08:17:58.375401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:17:58 crc kubenswrapper[4744]: I1201 08:17:58.375464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:17:59 crc kubenswrapper[4744]: I1201 08:17:59.534831 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:17:59 crc kubenswrapper[4744]: I1201 08:17:59.534951 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.134965 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.135282 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.136987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.137086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.137124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.229818 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.234850 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 08:18:00 crc kubenswrapper[4744]: I1201 08:18:00.234923 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 08:18:00 crc kubenswrapper[4744]: E1201 08:18:00.291070 4744 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 08:18:00 crc kubenswrapper[4744]: E1201 08:18:00.986941 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187d097e927cdffe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:17:48.227289086 +0000 UTC m=+0.216347017,LastTimestamp:2025-12-01 08:17:48.227289086 +0000 UTC m=+0.216347017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.162503 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.162588 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.172332 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.172446 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.988683 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.988849 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.995835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.995897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:01 crc kubenswrapper[4744]: I1201 08:18:01.995929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.072838 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.073014 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.074622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.074652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.074664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.075602 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.075772 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.076765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.076791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.076800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.081114 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.094860 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.391020 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.391033 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.392545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.392582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.392607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.392629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.392632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.392656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.435218 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 01 08:18:04 crc kubenswrapper[4744]: I1201 08:18:04.453597 4744 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.162600 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.165335 4744 trace.go:236] Trace[1582740734]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:17:51.374) (total time: 14790ms): Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[1582740734]: ---"Objects listed" error: 14790ms (08:18:06.165) Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[1582740734]: [14.790374058s] [14.790374058s] END Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.165753 4744 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.167133 4744 trace.go:236] Trace[1486626065]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:17:52.430) (total time: 13736ms): Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[1486626065]: ---"Objects listed" error: 13736ms (08:18:06.166) Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[1486626065]: [13.736139369s] [13.736139369s] END Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.167189 4744 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.168467 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.169964 4744 trace.go:236] Trace[559677149]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:17:51.416) (total time: 14752ms): Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[559677149]: ---"Objects listed" error: 14752ms (08:18:06.169) Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[559677149]: [14.752970509s] [14.752970509s] END Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.170001 4744 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.171845 4744 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.179750 4744 trace.go:236] Trace[496235207]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:17:52.699) (total time: 13480ms): Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[496235207]: ---"Objects listed" error: 13480ms (08:18:06.179) Dec 01 08:18:06 crc kubenswrapper[4744]: Trace[496235207]: [13.480246168s] [13.480246168s] END Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.179799 4744 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.208022 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38548->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.208080 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38564->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.208096 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38548->192.168.126.11:17697: read: connection reset by peer" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.208207 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38564->192.168.126.11:17697: read: connection reset by peer" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.208849 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.208916 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.234078 4744 apiserver.go:52] "Watching apiserver" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.256298 4744 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.257105 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.257677 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.257779 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.257799 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.257880 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.257912 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.258381 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.258453 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.258954 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.258461 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.260380 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.260717 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.262335 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.262767 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.262974 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.265020 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.265785 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.266505 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.267096 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.305115 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.327449 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.331091 4744 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.339377 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.349778 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.363062 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.373989 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374034 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374060 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374080 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374102 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374139 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374158 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374177 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374213 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374234 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374256 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374276 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374297 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374336 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374397 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374392 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374444 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374503 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374529 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374565 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374588 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374611 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374637 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374648 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374662 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374685 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374714 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374740 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374760 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374782 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374806 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374827 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374844 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374853 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374910 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374940 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374962 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.374985 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375011 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375031 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375055 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375077 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375099 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375144 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375154 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375167 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375192 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375215 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375238 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375261 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375286 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375336 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375363 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375386 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375433 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375458 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375480 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375502 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375524 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375545 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375568 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375591 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375616 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.375638 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.376638 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.383198 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.383366 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.383970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.384098 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.384242 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.384286 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.384499 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.384465 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.384666 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.384978 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.385045 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.385253 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.385305 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.385360 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.385791 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.387688 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.388259 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.388329 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.388886 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.388912 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.389053 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.389959 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.390013 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.390395 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.390671 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.390677 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.390741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.391309 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.391539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.391823 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.391990 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.392203 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.392462 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.392796 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.390359 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.393472 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.393499 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.394108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.394287 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.394835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.394923 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.391515 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.393577 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.395181 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.395357 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.395378 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.395566 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.396339 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.399085 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.395940 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.399511 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.397187 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.395235 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.399923 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.400342 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.400455 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.400928 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.400983 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401031 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401779 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401803 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401846 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401861 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401904 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401930 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401952 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401972 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.401992 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402070 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402088 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402108 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402160 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402172 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402187 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402180 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402243 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402269 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402293 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402317 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402338 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402417 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402439 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402452 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402461 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402505 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402530 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402610 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402639 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402638 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402663 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402688 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402711 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402736 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.402921 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.403323 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.403353 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.403711 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.403780 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404035 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404117 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404140 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404161 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404185 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404208 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404231 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404243 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404289 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404317 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404457 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.404859 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405324 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405338 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405660 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405712 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405735 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405763 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405770 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405793 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405815 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405837 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405860 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405886 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405910 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405935 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.405988 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406103 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406574 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406630 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406707 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406735 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406759 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406787 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406810 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406837 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406861 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406891 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406916 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406939 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406984 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407006 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407028 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407048 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407127 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407148 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407172 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407195 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407218 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407240 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407263 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407287 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407308 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407332 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407363 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407386 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407434 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407522 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407570 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407597 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407633 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407657 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407679 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407705 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407729 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407758 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407781 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407826 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407852 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407881 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407911 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407937 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407962 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407988 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408012 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408034 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408058 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408081 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408103 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408127 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408149 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408172 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408196 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408217 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408238 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408265 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408290 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408398 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408448 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408501 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408527 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408627 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408666 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408688 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408721 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408782 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408852 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408921 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408945 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408988 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409148 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409165 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409178 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409193 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409206 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409218 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409230 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409241 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409254 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409267 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409280 4744 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409293 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409305 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409320 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409335 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409354 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409368 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409380 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409393 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409421 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409436 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409449 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409461 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409474 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409486 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409499 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409516 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409529 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409544 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414053 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416197 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416274 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416294 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416309 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416324 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416375 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416392 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416419 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416436 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416457 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416472 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416487 4744 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416501 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416514 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416529 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416541 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416554 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416568 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416583 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416597 4744 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416611 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416624 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416638 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416652 4744 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416687 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416702 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416715 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416728 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416741 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416756 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416769 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416780 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416791 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416803 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416814 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416827 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416906 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416923 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416936 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416949 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416962 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416976 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417025 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417041 4744 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417054 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417090 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417104 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417117 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417129 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.419626 4744 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.420217 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406808 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.406833 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407081 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407257 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407323 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407683 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.407767 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408158 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408165 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408465 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.408900 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409067 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409315 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.409859 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.410243 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.410971 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.411253 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.411593 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.411700 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.412273 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.412360 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.412432 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.412821 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.413066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.413105 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.413380 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.413552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.413698 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414100 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414138 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.426494 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414576 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414665 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414833 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.414973 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.415112 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.429759 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.427921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.426647 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.415653 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.425824 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.415654 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416444 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416703 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.416922 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.418127 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.418589 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.427585 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.417814 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.418698 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.418713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.418795 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.419126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.419341 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.419712 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.419905 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.420369 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.420557 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.420814 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.421257 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.421505 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.421692 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.421989 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.422147 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.422694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.422823 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423019 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423248 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423389 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423398 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423493 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423731 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423874 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423894 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.423298 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.424795 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.425053 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.425084 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.425449 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:18:06.925385017 +0000 UTC m=+18.914442938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.430795 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:06.93076855 +0000 UTC m=+18.919826471 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.425472 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.425729 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.426072 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.426287 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.430929 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:06.930907164 +0000 UTC m=+18.919965085 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.426786 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.426991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.427057 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.427133 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.427255 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.427301 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.427556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.428113 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.428128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.433714 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.434556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.434878 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.435000 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.436601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.439624 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.439803 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902" exitCode=255 Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.439847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902"} Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.440504 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.440536 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.440553 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.440641 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:06.94061849 +0000 UTC m=+18.929676421 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.440939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.442131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.443319 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.443351 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.443841 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.444237 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.445571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.446869 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.447202 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.447754 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.447843 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.447874 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.448244 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.448292 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.448309 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.448379 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:06.948358441 +0000 UTC m=+18.937416462 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.449031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.449607 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.449816 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.450012 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.450077 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.450175 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.450373 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.450453 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.450533 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.450578 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.451120 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.452208 4744 scope.go:117] "RemoveContainer" containerID="d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.452620 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.456849 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.457112 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.458694 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.459942 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.465610 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.470919 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.481535 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.488129 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.498592 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.499601 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.509546 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525814 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525856 4744 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525866 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525875 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525885 4744 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525897 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525906 4744 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525914 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525922 4744 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525930 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525939 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525947 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525955 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525963 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525971 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525982 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.525993 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526006 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526015 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526024 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526032 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526039 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526048 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526058 4744 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526066 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526074 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526082 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526089 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526096 4744 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526104 4744 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526112 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526120 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526127 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526135 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526143 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526150 4744 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526158 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526165 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526172 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526180 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526187 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526196 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526203 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526211 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526218 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526225 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526232 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526240 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526248 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526255 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526262 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526270 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526277 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526285 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526294 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526301 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526309 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526318 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526326 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526334 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526342 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526351 4744 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526358 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526365 4744 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526373 4744 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526380 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526389 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526397 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526418 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526426 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526434 4744 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526442 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526449 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526456 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526464 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526471 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526480 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526490 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526498 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526507 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526514 4744 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526522 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526529 4744 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526539 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526547 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526555 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526563 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526571 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526579 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526587 4744 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526595 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526603 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526610 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526618 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526630 4744 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526638 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526647 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526655 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526662 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526671 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526678 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526687 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526695 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526703 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526710 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526718 4744 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526726 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526733 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526741 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526749 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526757 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526764 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526771 4744 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526780 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526795 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526803 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526811 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.526969 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.544780 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.544843 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.549063 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.553661 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.559068 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.572855 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.583318 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.585311 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.593037 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: W1201 08:18:06.601385 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-263d9016fc19f04a349a27cefbe949ae892fdd02d0b3d645d714ab0daeeb5882 WatchSource:0}: Error finding container 263d9016fc19f04a349a27cefbe949ae892fdd02d0b3d645d714ab0daeeb5882: Status 404 returned error can't find the container with id 263d9016fc19f04a349a27cefbe949ae892fdd02d0b3d645d714ab0daeeb5882 Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.602329 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.604747 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.616373 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.618800 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: W1201 08:18:06.619674 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-b449e9e54fa9fd56f9f8d7843657966ee517b328b0e2b549a5819203dd8ae5cf WatchSource:0}: Error finding container b449e9e54fa9fd56f9f8d7843657966ee517b328b0e2b549a5819203dd8ae5cf: Status 404 returned error can't find the container with id b449e9e54fa9fd56f9f8d7843657966ee517b328b0e2b549a5819203dd8ae5cf Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.629784 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.639902 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.649429 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.660361 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.671630 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.683014 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.693670 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.704509 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.714798 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:18:06 crc kubenswrapper[4744]: I1201 08:18:06.929563 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:06 crc kubenswrapper[4744]: E1201 08:18:06.929717 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:18:07.929702493 +0000 UTC m=+19.918760414 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.031125 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.031194 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.031218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.031242 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031336 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031386 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031473 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:08.031447499 +0000 UTC m=+20.020505420 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031509 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:08.03148301 +0000 UTC m=+20.020541151 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031592 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031606 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031624 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031639 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031669 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:08.031661745 +0000 UTC m=+20.020719666 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031608 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031690 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.031720 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:08.031714007 +0000 UTC m=+20.020771928 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.446997 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab"} Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.447080 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077"} Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.447106 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b449e9e54fa9fd56f9f8d7843657966ee517b328b0e2b549a5819203dd8ae5cf"} Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.450046 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e"} Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.450127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"263d9016fc19f04a349a27cefbe949ae892fdd02d0b3d645d714ab0daeeb5882"} Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.453213 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.459925 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0"} Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.460315 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.465855 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8706ec5e55463df9efe2ab76600d26ee5c3300bead6d0fb3ed0d468d638ab48a"} Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.479788 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.501494 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.524068 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.547351 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.565449 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.579616 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.594130 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.609012 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.623482 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.647618 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.659993 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.673173 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.684910 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.698525 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.711471 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.727081 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:07Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:07 crc kubenswrapper[4744]: I1201 08:18:07.939486 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:07 crc kubenswrapper[4744]: E1201 08:18:07.939711 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:18:09.939677415 +0000 UTC m=+21.928735346 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.040750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.040820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.040860 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.040892 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.040932 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041020 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:10.040998729 +0000 UTC m=+22.030056670 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041023 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041032 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041093 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041116 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041132 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:10.041112423 +0000 UTC m=+22.030170344 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041057 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041173 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041181 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041227 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:10.041217316 +0000 UTC m=+22.030275247 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.041244 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:10.041236686 +0000 UTC m=+22.030294627 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.284787 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.284878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.284996 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.285012 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.285143 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:08 crc kubenswrapper[4744]: E1201 08:18:08.285168 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.289700 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.290294 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.292068 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.292799 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.293941 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.294500 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.295111 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.296174 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.296801 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.297803 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.298326 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.299471 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.299995 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.300511 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.301495 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.302023 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.301992 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.302976 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.303355 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.303929 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.304938 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.305435 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.306449 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.306882 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.307898 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.308310 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.308920 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.310032 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.310523 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.311521 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.312084 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.313018 4744 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.313125 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.314970 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.315932 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.316348 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.318552 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.319221 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.320122 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.320810 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.321994 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.322464 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.323529 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.324207 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.325366 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.326291 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.326696 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.327217 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.327744 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.328875 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.329376 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.330528 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.331113 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.332138 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.332758 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.333257 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.343837 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.361643 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.385302 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.402943 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.424269 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:08 crc kubenswrapper[4744]: I1201 08:18:08.443922 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.369244 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.371003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.371034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.371046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.371147 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.379508 4744 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.380012 4744 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.382233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.382267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.382301 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.382318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.382331 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: E1201 08:18:09.411680 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.416930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.416986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.417003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.417024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.417057 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: E1201 08:18:09.435221 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.439443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.439512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.439535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.439582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.439600 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: E1201 08:18:09.453840 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.458359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.458397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.458425 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.458443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.458453 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: E1201 08:18:09.476665 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.481374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.481427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.481440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.481453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.481465 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: E1201 08:18:09.494208 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:09Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:09 crc kubenswrapper[4744]: E1201 08:18:09.494317 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.495761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.495783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.495791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.495803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.495815 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.598666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.598736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.598750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.598806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.598820 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.701524 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.701602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.701626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.701688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.701716 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.805797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.805897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.805921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.805948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.805966 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.909599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.909672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.909692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.909720 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.909740 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:09Z","lastTransitionTime":"2025-12-01T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:09 crc kubenswrapper[4744]: I1201 08:18:09.957336 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:09 crc kubenswrapper[4744]: E1201 08:18:09.957557 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:18:13.957530218 +0000 UTC m=+25.946588139 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.012601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.012673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.012695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.012721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.012739 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.058578 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.058665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.058727 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.058785 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.058901 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.058917 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059012 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:14.058987336 +0000 UTC m=+26.048045287 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059023 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059064 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059085 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059087 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059114 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059116 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059092 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:14.059059538 +0000 UTC m=+26.048117499 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059319 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:14.059260044 +0000 UTC m=+26.048318035 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.059378 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:14.059359377 +0000 UTC m=+26.048417418 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.116578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.116684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.116706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.116767 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.116792 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.220324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.220388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.220444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.220476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.220500 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.284685 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.284881 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.284704 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.284997 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.284686 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:10 crc kubenswrapper[4744]: E1201 08:18:10.285086 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.323556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.323646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.323664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.323687 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.323704 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.426358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.426438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.426452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.426471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.426483 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.478384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.503711 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.524371 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.530018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.530091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.530111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.530138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.530156 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.544663 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.578023 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.598305 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.619530 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.632787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.632868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.632892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.632921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.632944 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.639323 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.660917 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.736279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.736346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.736365 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.736396 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.736441 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.838958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.839010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.839024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.839041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.839053 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.941575 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.941640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.941662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.941688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:10 crc kubenswrapper[4744]: I1201 08:18:10.941706 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:10Z","lastTransitionTime":"2025-12-01T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.045228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.045308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.045336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.045369 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.045393 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.148781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.148834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.148850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.148872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.148886 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.251295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.251355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.251372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.251398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.251461 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.353999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.354341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.354353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.354370 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.354381 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.457181 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.457237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.457254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.457281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.457303 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.560718 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.560783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.560803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.560828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.560847 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.663587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.663635 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.663646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.663667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.663685 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.766460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.766531 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.766549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.766575 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.766591 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.868938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.868988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.869001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.869021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.869032 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.971931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.972019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.972035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.972058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:11 crc kubenswrapper[4744]: I1201 08:18:11.972073 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:11Z","lastTransitionTime":"2025-12-01T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.074513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.074543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.074553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.074585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.074594 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.178724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.178823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.179315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.179386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.179770 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.283216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.283294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.283314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.283733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.284013 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.284087 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.284096 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:12 crc kubenswrapper[4744]: E1201 08:18:12.284187 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:12 crc kubenswrapper[4744]: E1201 08:18:12.284257 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:12 crc kubenswrapper[4744]: E1201 08:18:12.284329 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.283758 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.387611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.387755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.387812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.387849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.387869 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.490235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.490320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.490344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.490374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.490397 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.593688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.593759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.593782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.593813 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.593834 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.697176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.697257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.697274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.697298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.697317 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.800939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.800989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.801004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.801023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.801038 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.904482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.904568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.904591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.904622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:12 crc kubenswrapper[4744]: I1201 08:18:12.904648 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:12Z","lastTransitionTime":"2025-12-01T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.007750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.007831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.007846 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.007868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.007937 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.110262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.110319 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.110332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.110348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.110377 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.137789 4744 csr.go:261] certificate signing request csr-4mx2q is approved, waiting to be issued Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.155085 4744 csr.go:257] certificate signing request csr-4mx2q is issued Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.212742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.212780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.212788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.212803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.212812 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.314876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.314932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.314941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.314956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.314970 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.417440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.417479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.417487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.417499 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.417508 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.520351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.520421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.520433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.520452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.520466 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.612947 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2cvbp"] Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.613371 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vsh4r"] Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.613539 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-ncr54"] Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.613584 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.613955 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.614014 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.615376 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.616350 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 08:18:13 crc kubenswrapper[4744]: W1201 08:18:13.616458 4744 reflector.go:561] object-"openshift-dns"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 01 08:18:13 crc kubenswrapper[4744]: E1201 08:18:13.616501 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.618272 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 08:18:13 crc kubenswrapper[4744]: W1201 08:18:13.618388 4744 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 01 08:18:13 crc kubenswrapper[4744]: E1201 08:18:13.618428 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.618526 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 08:18:13 crc kubenswrapper[4744]: W1201 08:18:13.620836 4744 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 01 08:18:13 crc kubenswrapper[4744]: E1201 08:18:13.620860 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:18:13 crc kubenswrapper[4744]: W1201 08:18:13.620924 4744 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 01 08:18:13 crc kubenswrapper[4744]: E1201 08:18:13.620936 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:18:13 crc kubenswrapper[4744]: W1201 08:18:13.620973 4744 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.621121 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 08:18:13 crc kubenswrapper[4744]: E1201 08:18:13.621319 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.623512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.623552 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.623562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.623580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.623594 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.632498 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.646958 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.658392 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.675333 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nxrc\" (UniqueName: \"kubernetes.io/projected/6a5e5532-8675-4f3f-aa30-7c22bc95b311-kube-api-access-7nxrc\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cnibin\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693495 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-conf-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-system-cni-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693575 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-k8s-cni-cncf-io\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693606 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-kubelet\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-etc-kubernetes\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693648 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-multus-certs\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693681 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-socket-dir-parent\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693714 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-netns\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/199cd5a5-cb9c-4e04-b7fb-da877b86eafc-hosts-file\") pod \"node-resolver-ncr54\" (UID: \"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\") " pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-os-release\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693786 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a5e5532-8675-4f3f-aa30-7c22bc95b311-cni-binary-copy\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693800 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-daemon-config\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmm4s\" (UniqueName: \"kubernetes.io/projected/199cd5a5-cb9c-4e04-b7fb-da877b86eafc-kube-api-access-zmm4s\") pod \"node-resolver-ncr54\" (UID: \"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\") " pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693876 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-system-cni-dir\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693912 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-cni-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693927 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-cnibin\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-cni-bin\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.693958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsm67\" (UniqueName: \"kubernetes.io/projected/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-kube-api-access-qsm67\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.694003 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-hostroot\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.694020 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.694044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-cni-multus\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.694063 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-os-release\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.694160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cni-binary-copy\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.699579 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.712477 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.723962 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.725349 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.725388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.725400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.725429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.725438 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.735271 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.745937 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.757141 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.769811 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.785449 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-os-release\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cni-binary-copy\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795160 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nxrc\" (UniqueName: \"kubernetes.io/projected/6a5e5532-8675-4f3f-aa30-7c22bc95b311-kube-api-access-7nxrc\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cnibin\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795237 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-conf-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-k8s-cni-cncf-io\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795256 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-os-release\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795270 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-system-cni-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-conf-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795305 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-system-cni-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795340 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-kubelet\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795390 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cnibin\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795456 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-k8s-cni-cncf-io\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-etc-kubernetes\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795539 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-etc-kubernetes\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-kubelet\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-multus-certs\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795586 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-multus-certs\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-socket-dir-parent\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-netns\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-os-release\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795793 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/199cd5a5-cb9c-4e04-b7fb-da877b86eafc-hosts-file\") pod \"node-resolver-ncr54\" (UID: \"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\") " pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795812 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-run-netns\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795814 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a5e5532-8675-4f3f-aa30-7c22bc95b311-cni-binary-copy\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-daemon-config\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/199cd5a5-cb9c-4e04-b7fb-da877b86eafc-hosts-file\") pod \"node-resolver-ncr54\" (UID: \"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\") " pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795881 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-system-cni-dir\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795908 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-socket-dir-parent\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmm4s\" (UniqueName: \"kubernetes.io/projected/199cd5a5-cb9c-4e04-b7fb-da877b86eafc-kube-api-access-zmm4s\") pod \"node-resolver-ncr54\" (UID: \"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\") " pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795942 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-system-cni-dir\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795953 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-cnibin\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795999 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-cni-bin\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsm67\" (UniqueName: \"kubernetes.io/projected/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-kube-api-access-qsm67\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796055 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-cni-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796072 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-hostroot\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796090 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796111 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-cni-multus\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796147 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-hostroot\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796175 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-cni-multus\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-host-var-lib-cni-bin\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795995 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-cnibin\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796272 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-cni-dir\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cni-binary-copy\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.795892 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-os-release\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a5e5532-8675-4f3f-aa30-7c22bc95b311-cni-binary-copy\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.796830 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.802781 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.820288 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.825725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nxrc\" (UniqueName: \"kubernetes.io/projected/6a5e5532-8675-4f3f-aa30-7c22bc95b311-kube-api-access-7nxrc\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.827571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.827608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.827619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.827638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.827649 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.832186 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsm67\" (UniqueName: \"kubernetes.io/projected/dc9dd3b9-b551-4a4b-8c4f-12ff6309034c-kube-api-access-qsm67\") pod \"multus-additional-cni-plugins-2cvbp\" (UID: \"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\") " pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.839132 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.850070 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.864202 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.878784 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.891246 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.906291 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.929185 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.929810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.929857 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.929869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.929887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.930217 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:13Z","lastTransitionTime":"2025-12-01T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:13 crc kubenswrapper[4744]: I1201 08:18:13.997639 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:13 crc kubenswrapper[4744]: E1201 08:18:13.997806 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:18:21.997782305 +0000 UTC m=+33.986840216 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.022721 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-g75sq"] Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.023171 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.025386 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.025878 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.026971 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.027144 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.034048 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.037015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.037047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.037055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.037070 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.037078 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.041750 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2sflk"] Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.050881 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.053117 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.053436 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.053666 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.053972 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.054155 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.054319 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.054472 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.056360 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.065639 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.082559 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-ovn\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-node-log\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-systemd-units\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098359 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-ovn-kubernetes\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098390 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-slash\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098438 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-config\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098471 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098492 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-systemd\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-var-lib-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098534 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098556 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098612 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh7zk\" (UniqueName: \"kubernetes.io/projected/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-kube-api-access-lh7zk\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098634 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1df124eb-0511-4702-85d8-3e324c59bb0d-ovn-node-metrics-cert\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098656 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-script-lib\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098675 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxzmp\" (UniqueName: \"kubernetes.io/projected/1df124eb-0511-4702-85d8-3e324c59bb0d-kube-api-access-hxzmp\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-rootfs\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-bin\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-env-overrides\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-netns\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-netd\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-mcd-auth-proxy-config\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098855 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-proxy-tls\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098874 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-kubelet\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098893 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-etc-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.098921 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-log-socket\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.099057 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.099107 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:22.099092289 +0000 UTC m=+34.088150210 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.103757 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.103785 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.103798 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.103837 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:22.103824054 +0000 UTC m=+34.092881975 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.103906 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.103935 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:22.103924107 +0000 UTC m=+34.092982028 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.104076 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.104094 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.104102 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.104126 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:22.104119042 +0000 UTC m=+34.093176973 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.104776 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.126369 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.138748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.138782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.138795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.138812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.138821 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.138911 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.153324 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.156353 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-01 08:13:13 +0000 UTC, rotation deadline is 2026-08-18 03:40:24.429356354 +0000 UTC Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.156396 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6235h22m10.272961714s for next certificate rotation Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.165526 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.177161 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.193232 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.199901 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-etc-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.199936 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-log-socket\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.199959 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-ovn\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.199974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-node-log\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.199990 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-systemd-units\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200013 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-ovn-kubernetes\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-slash\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200058 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-config\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-systemd\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200110 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-var-lib-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200135 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh7zk\" (UniqueName: \"kubernetes.io/projected/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-kube-api-access-lh7zk\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-node-log\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200167 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1df124eb-0511-4702-85d8-3e324c59bb0d-ovn-node-metrics-cert\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-script-lib\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxzmp\" (UniqueName: \"kubernetes.io/projected/1df124eb-0511-4702-85d8-3e324c59bb0d-kube-api-access-hxzmp\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200305 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-rootfs\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-bin\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200358 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-env-overrides\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200388 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-netns\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200420 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-netd\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-mcd-auth-proxy-config\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200467 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-kubelet\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.200493 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-proxy-tls\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-bin\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201080 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-etc-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201102 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-ovn-kubernetes\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-log-socket\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201140 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-var-lib-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201155 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-ovn\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201146 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-systemd-units\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201196 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-rootfs\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201200 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-systemd\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-netd\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-netns\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201321 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-kubelet\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201347 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-slash\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.201372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-openvswitch\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.202181 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-mcd-auth-proxy-config\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.202252 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-config\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.204005 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-proxy-tls\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.204347 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-env-overrides\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.206144 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-script-lib\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.206646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1df124eb-0511-4702-85d8-3e324c59bb0d-ovn-node-metrics-cert\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.209531 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.218521 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxzmp\" (UniqueName: \"kubernetes.io/projected/1df124eb-0511-4702-85d8-3e324c59bb0d-kube-api-access-hxzmp\") pod \"ovnkube-node-2sflk\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.219133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh7zk\" (UniqueName: \"kubernetes.io/projected/c8ecf18a-4cfa-4266-87ae-1a08097cb5d1-kube-api-access-lh7zk\") pod \"machine-config-daemon-g75sq\" (UID: \"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\") " pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.221991 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.238213 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.241055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.241089 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.241098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.241113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.241124 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.256015 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.266744 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.275133 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.284127 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.284182 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.284129 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.284238 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.284311 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:14 crc kubenswrapper[4744]: E1201 08:18:14.284375 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.289050 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.297120 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.307430 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.319164 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.332125 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.334966 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.346779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.346829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.346843 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.346862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.346880 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.348752 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.364705 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.380172 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.393027 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.415197 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:14 crc kubenswrapper[4744]: W1201 08:18:14.428055 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1df124eb_0511_4702_85d8_3e324c59bb0d.slice/crio-e4f00f2099e951bf715b1b3196e6597595c7cefbabb3eb9f71ccbaa0ef227d6c WatchSource:0}: Error finding container e4f00f2099e951bf715b1b3196e6597595c7cefbabb3eb9f71ccbaa0ef227d6c: Status 404 returned error can't find the container with id e4f00f2099e951bf715b1b3196e6597595c7cefbabb3eb9f71ccbaa0ef227d6c Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.449137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.449179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.449188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.449202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.449212 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.464910 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.496072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"e4f00f2099e951bf715b1b3196e6597595c7cefbabb3eb9f71ccbaa0ef227d6c"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.497989 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.498035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"d63c5a8c4714cb5971fbddb9abfd80ffcf84193e882232ce73f053d5bcab9d6b"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.500155 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc9dd3b9-b551-4a4b-8c4f-12ff6309034c" containerID="6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa" exitCode=0 Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.500198 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerDied","Data":"6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.500221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerStarted","Data":"d7b02ab1dcc7709bf36850bcfe4a5a585ae64df0b9d2eea643c418b45fb5e514"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.518463 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.537654 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.539590 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.552391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.552442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.552453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.552490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.552501 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.555708 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.568353 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.586188 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.596600 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.610054 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.630197 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.643710 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.652638 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.654779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.654812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.654827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.654845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.654857 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.656041 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.669500 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.670573 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.676774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a5e5532-8675-4f3f-aa30-7c22bc95b311-multus-daemon-config\") pod \"multus-vsh4r\" (UID: \"6a5e5532-8675-4f3f-aa30-7c22bc95b311\") " pod="openshift-multus/multus-vsh4r" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.681240 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.694961 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:14Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.757526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.757568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.757579 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.757595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.757607 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.841023 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vsh4r" Dec 01 08:18:14 crc kubenswrapper[4744]: W1201 08:18:14.853259 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a5e5532_8675_4f3f_aa30_7c22bc95b311.slice/crio-f535e4205932124d6ceefdff154c8ccc769e7ae7288f9137587bf52aa02c86ff WatchSource:0}: Error finding container f535e4205932124d6ceefdff154c8ccc769e7ae7288f9137587bf52aa02c86ff: Status 404 returned error can't find the container with id f535e4205932124d6ceefdff154c8ccc769e7ae7288f9137587bf52aa02c86ff Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.860883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.860928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.860947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.860985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.861004 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.892878 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.903610 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmm4s\" (UniqueName: \"kubernetes.io/projected/199cd5a5-cb9c-4e04-b7fb-da877b86eafc-kube-api-access-zmm4s\") pod \"node-resolver-ncr54\" (UID: \"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\") " pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.964316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.964348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.964358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.964377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:14 crc kubenswrapper[4744]: I1201 08:18:14.964387 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:14Z","lastTransitionTime":"2025-12-01T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.066426 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.066464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.066473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.066494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.066505 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.134345 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ncr54" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.180214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.180644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.180654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.180671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.180704 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.283921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.283988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.284006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.284031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.284049 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.386376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.386436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.386451 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.386465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.386476 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.489997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.490060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.490072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.490096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.490113 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.505542 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6" exitCode=0 Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.505657 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.510150 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.512979 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc9dd3b9-b551-4a4b-8c4f-12ff6309034c" containerID="7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075" exitCode=0 Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.513055 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerDied","Data":"7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.514271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ncr54" event={"ID":"199cd5a5-cb9c-4e04-b7fb-da877b86eafc","Type":"ContainerStarted","Data":"dffc83aea7eeafbfb4d167935b309b0ad35fe985eff8ab3aa2c40ec02b73bf64"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.515218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerStarted","Data":"f535e4205932124d6ceefdff154c8ccc769e7ae7288f9137587bf52aa02c86ff"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.527909 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.544558 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-z4dh2"] Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.544898 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.545246 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.550568 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.550732 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.550898 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.551062 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.564338 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.579116 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.592158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.592188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.592196 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.592211 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.592221 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.595785 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.616101 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.616487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmlxw\" (UniqueName: \"kubernetes.io/projected/cac4540b-1bb4-43e1-b373-c0bff223afdd-kube-api-access-fmlxw\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.616615 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cac4540b-1bb4-43e1-b373-c0bff223afdd-host\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.616684 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cac4540b-1bb4-43e1-b373-c0bff223afdd-serviceca\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.636315 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.649060 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.663997 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.680665 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.692433 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.704194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.704258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.704274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.704295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.704313 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.706299 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.717429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmlxw\" (UniqueName: \"kubernetes.io/projected/cac4540b-1bb4-43e1-b373-c0bff223afdd-kube-api-access-fmlxw\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.717475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cac4540b-1bb4-43e1-b373-c0bff223afdd-host\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.717510 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cac4540b-1bb4-43e1-b373-c0bff223afdd-serviceca\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.717596 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cac4540b-1bb4-43e1-b373-c0bff223afdd-host\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.718588 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cac4540b-1bb4-43e1-b373-c0bff223afdd-serviceca\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.724464 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.737589 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.738568 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmlxw\" (UniqueName: \"kubernetes.io/projected/cac4540b-1bb4-43e1-b373-c0bff223afdd-kube-api-access-fmlxw\") pod \"node-ca-z4dh2\" (UID: \"cac4540b-1bb4-43e1-b373-c0bff223afdd\") " pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.752617 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.768678 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.781279 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.796681 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.807799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.807839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.807851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.807867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.807879 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.811550 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.821931 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.847656 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.868440 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.879758 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.893720 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-z4dh2" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.899134 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.910104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.910349 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.910358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.910371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.910384 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:15Z","lastTransitionTime":"2025-12-01T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.913381 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: W1201 08:18:15.917510 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcac4540b_1bb4_43e1_b373_c0bff223afdd.slice/crio-4abf5cfaba9110798db2a31496d6f07b68d937400a754c593e186d87b33f501b WatchSource:0}: Error finding container 4abf5cfaba9110798db2a31496d6f07b68d937400a754c593e186d87b33f501b: Status 404 returned error can't find the container with id 4abf5cfaba9110798db2a31496d6f07b68d937400a754c593e186d87b33f501b Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.935422 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:15 crc kubenswrapper[4744]: I1201 08:18:15.951162 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:15Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.013241 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.013291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.013303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.013323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.013336 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.116572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.116634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.116652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.116673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.116690 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.219114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.219170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.219186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.219208 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.219225 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.284130 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.284186 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.284268 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:16 crc kubenswrapper[4744]: E1201 08:18:16.284386 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:16 crc kubenswrapper[4744]: E1201 08:18:16.284575 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:16 crc kubenswrapper[4744]: E1201 08:18:16.284686 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.327157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.327210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.327221 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.327238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.327254 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.429484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.429539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.429553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.429573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.429585 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.527625 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc9dd3b9-b551-4a4b-8c4f-12ff6309034c" containerID="e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2" exitCode=0 Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.527710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerDied","Data":"e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.530184 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerStarted","Data":"49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.531717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.531765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.531783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.531808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.531826 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.532886 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-z4dh2" event={"ID":"cac4540b-1bb4-43e1-b373-c0bff223afdd","Type":"ContainerStarted","Data":"291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.532947 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-z4dh2" event={"ID":"cac4540b-1bb4-43e1-b373-c0bff223afdd","Type":"ContainerStarted","Data":"4abf5cfaba9110798db2a31496d6f07b68d937400a754c593e186d87b33f501b"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.536710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ncr54" event={"ID":"199cd5a5-cb9c-4e04-b7fb-da877b86eafc","Type":"ContainerStarted","Data":"d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.547054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.547120 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.547141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.547160 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.547178 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.547196 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.549302 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.565377 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.582776 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.600859 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.613135 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.623472 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.634227 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.634616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.634651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.634661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.634674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.634683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.646166 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.663450 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.684668 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.697085 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.709743 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.724244 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.737336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.737400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.737509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.737540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.737572 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.739234 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.794535 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.814271 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.833817 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.839387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.839454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.839489 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.839506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.839516 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.861908 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.876485 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.888378 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.900990 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.914925 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.942003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.942060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.942077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.942101 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.942115 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:16Z","lastTransitionTime":"2025-12-01T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.944572 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.954866 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.968610 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.980244 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:16 crc kubenswrapper[4744]: I1201 08:18:16.990371 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.000661 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.045121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.045156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.045170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.045187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.045200 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.148385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.148430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.148438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.148454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.148463 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.250480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.250512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.250520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.250535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.250544 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.353594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.354116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.354146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.354184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.354204 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.456883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.456917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.456925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.456941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.456952 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.552319 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc9dd3b9-b551-4a4b-8c4f-12ff6309034c" containerID="1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2" exitCode=0 Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.553058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerDied","Data":"1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.559333 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.559376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.559390 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.559433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.559448 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.578179 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.595218 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.610447 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.623105 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.636929 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.646997 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.658239 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.663232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.663270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.663283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.663303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.663316 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.669691 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.683347 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.703314 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.722058 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.734469 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.750382 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.761038 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:17Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.767821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.767867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.767880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.767901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.767914 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.870859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.870906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.870922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.870946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.870963 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.973389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.973436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.973447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.973463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:17 crc kubenswrapper[4744]: I1201 08:18:17.973471 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:17Z","lastTransitionTime":"2025-12-01T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.077011 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.077092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.077117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.077149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.077174 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.159617 4744 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.189740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.189819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.189836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.189895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.189912 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.283954 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:18 crc kubenswrapper[4744]: E1201 08:18:18.284260 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.284581 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.284648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:18 crc kubenswrapper[4744]: E1201 08:18:18.284728 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:18 crc kubenswrapper[4744]: E1201 08:18:18.284849 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.293641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.293701 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.293726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.293754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.293778 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.303041 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.325881 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.346403 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.363175 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.400265 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.401541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.401593 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.401610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.401633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.401651 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.418677 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.431371 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.451689 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.488297 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.501237 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.505099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.505125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.505135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.505152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.505163 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.524004 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.544846 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.561272 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc9dd3b9-b551-4a4b-8c4f-12ff6309034c" containerID="25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382" exitCode=0 Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.561328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerDied","Data":"25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.571718 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.573692 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.590882 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.605478 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.607308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.607351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.607363 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.607380 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.607390 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.622672 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.648437 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.661894 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.683247 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.697358 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.710074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.710110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.710119 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.710135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.710144 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.710065 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.724355 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.748444 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.760484 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.812481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.812518 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.812528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.812545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.812556 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.839999 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.853840 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.865525 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.880291 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.914815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.914864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.914916 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.914941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:18 crc kubenswrapper[4744]: I1201 08:18:18.914957 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:18Z","lastTransitionTime":"2025-12-01T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.017681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.017722 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.017738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.017757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.017774 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.121260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.121305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.121322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.121350 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.121369 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.224923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.225002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.225021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.225047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.225122 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.328808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.328858 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.328877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.328896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.328910 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.432458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.432690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.432708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.432741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.432765 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.536002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.536039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.536047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.536062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.536070 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.580132 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc9dd3b9-b551-4a4b-8c4f-12ff6309034c" containerID="dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe" exitCode=0 Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.580180 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerDied","Data":"dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.596039 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.629647 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.635782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.635825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.635836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.635854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.635869 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.645031 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: E1201 08:18:19.648003 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.653093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.653143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.653157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.653177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.653192 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.656257 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: E1201 08:18:19.665927 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.670762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.670810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.670830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.670736 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.670850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.670987 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: E1201 08:18:19.686053 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.688187 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.689684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.689732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.689745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.689765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.689780 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.703684 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: E1201 08:18:19.706124 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.714043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.714082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.714274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.714317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.714331 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.724097 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: E1201 08:18:19.727701 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: E1201 08:18:19.728045 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.733064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.733101 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.733111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.733129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.733143 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.741747 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.761106 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.776946 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.790793 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.812254 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.826973 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:19Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.835787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.835814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.835823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.835836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.835846 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.937331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.937362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.937369 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.937382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:19 crc kubenswrapper[4744]: I1201 08:18:19.937392 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:19Z","lastTransitionTime":"2025-12-01T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.040710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.040785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.040809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.040838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.040861 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.146838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.146914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.146940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.146972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.146995 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.242638 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.250135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.250190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.250207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.250229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.250247 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.270112 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.284054 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.284130 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.284072 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:20 crc kubenswrapper[4744]: E1201 08:18:20.284263 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:20 crc kubenswrapper[4744]: E1201 08:18:20.284382 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:20 crc kubenswrapper[4744]: E1201 08:18:20.284535 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.291756 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.309949 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.334038 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.351297 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.354976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.355026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.355042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.355065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.355078 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.370660 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.387122 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.407025 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.426580 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.449695 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.457952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.458005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.458021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.458047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.458063 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.472081 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.490614 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.505885 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.522139 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.561529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.561572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.561591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.561649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.561669 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.590646 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" event={"ID":"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c","Type":"ContainerStarted","Data":"a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.610322 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.629874 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.647175 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.664089 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.665988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.666168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.666192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.666288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.666366 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.685835 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.700067 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.718720 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.753786 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.768849 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.771121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.771179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.771201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.771233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.771256 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.787603 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.806963 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.827877 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.843852 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.866149 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.873400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.873458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.873474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.873494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.873508 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.976500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.976556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.976568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.976588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:20 crc kubenswrapper[4744]: I1201 08:18:20.976602 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:20Z","lastTransitionTime":"2025-12-01T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.080098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.080171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.080196 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.080230 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.080255 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.183853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.183915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.183933 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.183961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.183985 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.287235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.287292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.287310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.287343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.287368 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.390976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.391051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.391076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.391109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.391130 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.494811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.494874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.494890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.494914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.494930 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.597804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.597854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.597878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.597911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.597936 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.603055 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.603605 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.638180 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.642351 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.651907 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.664034 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.679399 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.696640 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.700736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.700792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.700815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.700851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.700878 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.718389 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.736331 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.753207 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.775730 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.790788 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.801614 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.803546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.803608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.803627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.803653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.803671 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.821252 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.844465 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.857086 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.879447 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.892919 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.905991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.906030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.906047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.906069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.906085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:21Z","lastTransitionTime":"2025-12-01T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.911292 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.930436 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.950520 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.964385 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.981727 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:21 crc kubenswrapper[4744]: I1201 08:18:21.997994 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.000003 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.000206 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:18:38.000189226 +0000 UTC m=+49.989247177 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.008549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.008802 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.008913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.009027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.009127 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.011116 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.027989 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.041429 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.061394 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.079980 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.099352 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.100761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.100895 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.100972 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:38.100952404 +0000 UTC m=+50.090010365 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.112577 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.112833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.113138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.113462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.113750 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.201770 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202027 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202071 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202094 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.202243 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202313 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202439 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202445 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:38.202349531 +0000 UTC m=+50.191407482 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202466 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.202351 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202600 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:38.202584657 +0000 UTC m=+50.191642608 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.202919 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.203065 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:38.203047731 +0000 UTC m=+50.192105662 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.216768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.216991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.217125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.217259 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.217446 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.284563 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.284563 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.284664 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.285502 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.285114 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:22 crc kubenswrapper[4744]: E1201 08:18:22.285914 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.321280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.321335 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.321374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.321433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.321451 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.424215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.424257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.424272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.424289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.424301 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.526404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.526503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.526528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.526557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.526581 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.607640 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.608404 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.629338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.629395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.629458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.629493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.629511 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.644279 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.660887 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.674872 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.697178 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.713464 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.727125 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.731985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.732067 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.732093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.732120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.732139 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.741763 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.775361 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.791090 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.809327 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.828043 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.835243 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.835308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.835327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.835354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.835372 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.848366 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.868852 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.889695 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.909931 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.938220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.938265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.938276 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.938294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:22 crc kubenswrapper[4744]: I1201 08:18:22.938309 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:22Z","lastTransitionTime":"2025-12-01T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.040926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.041106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.041114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.041126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.041134 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.143377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.143429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.143441 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.143460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.143473 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.245526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.245572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.245587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.245604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.245615 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.349676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.349730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.349747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.349770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.349790 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.452005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.452063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.452083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.452108 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.452125 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.555139 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.555221 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.555248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.555276 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.555295 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.610974 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.660919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.660999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.661021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.661051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.661073 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.764101 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.764157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.764176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.764199 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.764219 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.867060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.867114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.867132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.867157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.867173 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.969438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.969505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.969514 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.969528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:23 crc kubenswrapper[4744]: I1201 08:18:23.969537 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:23Z","lastTransitionTime":"2025-12-01T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.072992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.073104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.073124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.073146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.073161 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.176229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.176300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.176317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.176342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.176359 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.279504 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.279583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.279599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.279627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.279645 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.284933 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:24 crc kubenswrapper[4744]: E1201 08:18:24.285130 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.285801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:24 crc kubenswrapper[4744]: E1201 08:18:24.285937 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.287054 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:24 crc kubenswrapper[4744]: E1201 08:18:24.289554 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.383022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.383082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.383102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.383129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.383146 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.486334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.486389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.486446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.486478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.486504 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.588986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.589042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.589060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.589111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.589128 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.617827 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/0.log" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.622339 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc" exitCode=1 Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.622395 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.623436 4744 scope.go:117] "RemoveContainer" containerID="1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.655737 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:23Z\\\",\\\"message\\\":\\\"140\\\\nI1201 08:18:23.076497 6054 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:18:23.076541 6054 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:18:23.076542 6054 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:18:23.076552 6054 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:18:23.076565 6054 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:18:23.076631 6054 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:18:23.076637 6054 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.076754 6054 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.077083 6054 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:18:23.077105 6054 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:18:23.077131 6054 factory.go:656] Stopping watch factory\\\\nI1201 08:18:23.077156 6054 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:18:23.077183 6054 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.674050 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.693325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.693387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.693430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.693455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.693475 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.694160 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.711193 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.732941 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.752225 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.775891 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.796881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.796963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.796981 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.797012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.797035 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.797976 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.815741 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.833732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.853200 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.877137 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.899949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.900060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.900073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.900133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.900149 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:24Z","lastTransitionTime":"2025-12-01T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.901100 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:24 crc kubenswrapper[4744]: I1201 08:18:24.917549 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:24Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.003110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.003149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.003161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.003178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.003190 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.105915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.105957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.105968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.105987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.105998 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.208234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.208304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.208322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.208348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.208372 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.311081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.311145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.311160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.311183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.311203 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.413750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.413817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.413834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.413862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.413881 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.516370 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.516471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.516498 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.516531 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.516554 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.619548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.619633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.619658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.619692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.619753 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.627276 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/0.log" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.630996 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.631098 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.654722 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.669686 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.693437 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.709217 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.722829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.722901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.722917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.722944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.722963 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.732937 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.765208 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:23Z\\\",\\\"message\\\":\\\"140\\\\nI1201 08:18:23.076497 6054 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:18:23.076541 6054 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:18:23.076542 6054 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:18:23.076552 6054 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:18:23.076565 6054 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:18:23.076631 6054 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:18:23.076637 6054 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.076754 6054 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.077083 6054 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:18:23.077105 6054 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:18:23.077131 6054 factory.go:656] Stopping watch factory\\\\nI1201 08:18:23.077156 6054 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:18:23.077183 6054 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.780822 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.797287 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.811871 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.825477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.825517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.825539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.825558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.825572 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.827958 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.843476 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.860116 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.873762 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.887614 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:25Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.928332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.928393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.928437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.928468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:25 crc kubenswrapper[4744]: I1201 08:18:25.928482 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:25Z","lastTransitionTime":"2025-12-01T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.034704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.034969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.035587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.035624 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.035650 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.139053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.139134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.139151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.139177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.139196 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.242710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.242782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.242800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.242826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.242843 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.284594 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.284686 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:26 crc kubenswrapper[4744]: E1201 08:18:26.284730 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.284755 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:26 crc kubenswrapper[4744]: E1201 08:18:26.284893 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:26 crc kubenswrapper[4744]: E1201 08:18:26.284991 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.344867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.344922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.344937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.344959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.344973 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.447834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.447868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.447878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.447893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.447902 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.550342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.550474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.550500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.550526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.550544 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.636549 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/1.log" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.637145 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/0.log" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.641984 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a" exitCode=1 Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.642029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.642083 4744 scope.go:117] "RemoveContainer" containerID="1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.642813 4744 scope.go:117] "RemoveContainer" containerID="52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a" Dec 01 08:18:26 crc kubenswrapper[4744]: E1201 08:18:26.643031 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.653558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.653618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.653629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.653647 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.653660 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.658824 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.671574 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.686624 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.705318 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.721746 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.734105 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.749560 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.756320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.756390 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.756467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.756497 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.756515 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.762778 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.780161 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.793076 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.806758 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.827061 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:23Z\\\",\\\"message\\\":\\\"140\\\\nI1201 08:18:23.076497 6054 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:18:23.076541 6054 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:18:23.076542 6054 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:18:23.076552 6054 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:18:23.076565 6054 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:18:23.076631 6054 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:18:23.076637 6054 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.076754 6054 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.077083 6054 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:18:23.077105 6054 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:18:23.077131 6054 factory.go:656] Stopping watch factory\\\\nI1201 08:18:23.077156 6054 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:18:23.077183 6054 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.839210 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.852883 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.859215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.859275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.859298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.859321 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.859336 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.962822 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.962879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.962897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.962924 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:26 crc kubenswrapper[4744]: I1201 08:18:26.962942 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:26Z","lastTransitionTime":"2025-12-01T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.012487 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc"] Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.013669 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.016028 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.019160 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.037237 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.056604 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.065824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.065882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.065892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.065915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.065927 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.073506 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.095340 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.119760 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.138258 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.151439 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.153972 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/707e21d6-1ab8-40f2-8681-52a0263cddff-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.154069 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/707e21d6-1ab8-40f2-8681-52a0263cddff-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.154283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/707e21d6-1ab8-40f2-8681-52a0263cddff-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.154449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9d6j\" (UniqueName: \"kubernetes.io/projected/707e21d6-1ab8-40f2-8681-52a0263cddff-kube-api-access-s9d6j\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.168953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.168994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.169003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.169018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.169027 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.171271 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.200811 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1848d1cc2afe242d18610f35893938aebb35770c758d340345963334308001bc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:23Z\\\",\\\"message\\\":\\\"140\\\\nI1201 08:18:23.076497 6054 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:18:23.076541 6054 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:18:23.076542 6054 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:18:23.076552 6054 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 08:18:23.076565 6054 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:18:23.076631 6054 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 08:18:23.076637 6054 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.076754 6054 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:18:23.077083 6054 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:18:23.077105 6054 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:18:23.077131 6054 factory.go:656] Stopping watch factory\\\\nI1201 08:18:23.077156 6054 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:18:23.077183 6054 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.213255 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.232230 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.249819 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.255470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/707e21d6-1ab8-40f2-8681-52a0263cddff-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.255590 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/707e21d6-1ab8-40f2-8681-52a0263cddff-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.255644 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9d6j\" (UniqueName: \"kubernetes.io/projected/707e21d6-1ab8-40f2-8681-52a0263cddff-kube-api-access-s9d6j\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.255745 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/707e21d6-1ab8-40f2-8681-52a0263cddff-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.257021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/707e21d6-1ab8-40f2-8681-52a0263cddff-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.257065 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/707e21d6-1ab8-40f2-8681-52a0263cddff-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.264593 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/707e21d6-1ab8-40f2-8681-52a0263cddff-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.273531 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.273594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.273613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.273638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.273657 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.279022 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9d6j\" (UniqueName: \"kubernetes.io/projected/707e21d6-1ab8-40f2-8681-52a0263cddff-kube-api-access-s9d6j\") pod \"ovnkube-control-plane-749d76644c-lwkvc\" (UID: \"707e21d6-1ab8-40f2-8681-52a0263cddff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.279523 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.298306 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.320658 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.338912 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" Dec 01 08:18:27 crc kubenswrapper[4744]: W1201 08:18:27.364852 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod707e21d6_1ab8_40f2_8681_52a0263cddff.slice/crio-5d4d0ca4d05dad23feb6191b62a312c3d11aca24cb9d1b07ea056a10b5773066 WatchSource:0}: Error finding container 5d4d0ca4d05dad23feb6191b62a312c3d11aca24cb9d1b07ea056a10b5773066: Status 404 returned error can't find the container with id 5d4d0ca4d05dad23feb6191b62a312c3d11aca24cb9d1b07ea056a10b5773066 Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.395368 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.395459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.395479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.395505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.395522 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.497731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.497770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.497779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.497795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.497804 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.600596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.600638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.600651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.600669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.600683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.646494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" event={"ID":"707e21d6-1ab8-40f2-8681-52a0263cddff","Type":"ContainerStarted","Data":"5d4d0ca4d05dad23feb6191b62a312c3d11aca24cb9d1b07ea056a10b5773066"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.648997 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/1.log" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.654733 4744 scope.go:117] "RemoveContainer" containerID="52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a" Dec 01 08:18:27 crc kubenswrapper[4744]: E1201 08:18:27.654985 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.669657 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.689124 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.703649 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.705563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.705628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.705653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.705688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.705715 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.722241 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.750506 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.767781 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.788263 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.807048 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.809183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.809221 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.809234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.809251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.809262 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.827626 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.845940 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.864479 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.880000 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.904773 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.911119 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.911176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.911194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.911219 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.911237 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:27Z","lastTransitionTime":"2025-12-01T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.917267 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:27 crc kubenswrapper[4744]: I1201 08:18:27.930233 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:27Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.014495 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.014556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.014579 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.014608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.014629 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.117194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.117224 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.117232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.117247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.117709 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.181622 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-p489k"] Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.182297 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.182386 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.203570 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.220678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.220886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.221026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.221168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.221323 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.233991 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.256074 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.265587 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfnrb\" (UniqueName: \"kubernetes.io/projected/c8ac81c1-4bf7-472c-8110-21364dde7fd6-kube-api-access-zfnrb\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.265656 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.275108 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.284971 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.285080 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.285778 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.285160 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.286130 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.285109 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.286553 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.302121 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.314054 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.323671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.323703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.323715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.323731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.323742 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.327427 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.338724 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.351618 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.364722 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.366402 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.366480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfnrb\" (UniqueName: \"kubernetes.io/projected/c8ac81c1-4bf7-472c-8110-21364dde7fd6-kube-api-access-zfnrb\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.367045 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.367142 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:28.867115266 +0000 UTC m=+40.856173207 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.384326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfnrb\" (UniqueName: \"kubernetes.io/projected/c8ac81c1-4bf7-472c-8110-21364dde7fd6-kube-api-access-zfnrb\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.390356 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.407888 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.417681 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.426005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.426031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.426039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.426053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.426062 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.434704 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.445931 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.459823 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.474777 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.486520 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.495028 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.504383 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.520899 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.528325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.528371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.528384 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.528402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.528432 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.538274 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.551075 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.563567 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.578565 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.588355 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.603805 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.615664 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.632237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.632282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.632294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.632312 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.632326 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.632276 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.654382 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.659378 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" event={"ID":"707e21d6-1ab8-40f2-8681-52a0263cddff","Type":"ContainerStarted","Data":"3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.659497 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" event={"ID":"707e21d6-1ab8-40f2-8681-52a0263cddff","Type":"ContainerStarted","Data":"0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.667222 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.682623 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.702276 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.716375 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.729402 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.735096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.735135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.735146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.735163 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.735176 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.746909 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.759768 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.772568 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.783282 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.796712 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.809041 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.821752 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.838292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.838347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.838363 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.838388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.838428 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.838818 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.851259 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.867626 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.871665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.871865 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:28 crc kubenswrapper[4744]: E1201 08:18:28.871957 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:29.871936978 +0000 UTC m=+41.860994899 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.879475 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.894025 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.941188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.941226 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.941237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.941253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:28 crc kubenswrapper[4744]: I1201 08:18:28.941264 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:28Z","lastTransitionTime":"2025-12-01T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.044272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.044302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.044313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.044329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.044342 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.147314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.147445 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.147465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.147494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.147512 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.250372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.250475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.250515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.250547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.250564 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.354382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.354465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.354482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.354505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.354513 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.457569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.457911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.457929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.457956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.457973 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.561551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.561600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.561618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.561643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.561659 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.663967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.664048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.664074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.664105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.664126 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.767199 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.767285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.767308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.767333 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.767352 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.871008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.871061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.871082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.871111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.871132 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.881952 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:29 crc kubenswrapper[4744]: E1201 08:18:29.882110 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:29 crc kubenswrapper[4744]: E1201 08:18:29.882200 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:31.882174197 +0000 UTC m=+43.871232148 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.974469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.974545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.974564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.974589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.974609 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.983309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.983369 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.983386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.983435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:29 crc kubenswrapper[4744]: I1201 08:18:29.983454 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:29Z","lastTransitionTime":"2025-12-01T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.003293 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.008759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.008809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.008828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.008853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.008872 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.028914 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.033907 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.033994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.034016 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.034046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.034068 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.054287 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.058928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.058997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.059021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.059054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.059080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.079198 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.084030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.084084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.084103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.084132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.084159 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.103807 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:30Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.104071 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.106448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.106501 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.106520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.106544 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.106562 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.208602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.208683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.208710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.208741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.208759 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.284627 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.284735 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.284758 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.284832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.284877 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.285013 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.285144 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:30 crc kubenswrapper[4744]: E1201 08:18:30.285596 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.311774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.311873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.311891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.311918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.311940 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.415051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.415127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.415148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.415182 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.415210 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.518511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.518583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.518599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.518627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.518644 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.621554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.621611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.621628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.621654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.621672 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.724901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.725069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.725103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.725134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.725155 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.828496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.828565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.828582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.828608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.828628 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.931970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.932039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.932059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.932084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:30 crc kubenswrapper[4744]: I1201 08:18:30.932101 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:30Z","lastTransitionTime":"2025-12-01T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.035902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.035961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.035979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.036005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.036023 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.139242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.139324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.139347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.139376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.139398 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.243004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.243092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.243117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.243150 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.243174 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.350565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.350631 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.350653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.350694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.350719 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.453948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.453989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.453996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.454009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.454018 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.557342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.557433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.557452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.557477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.557497 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.660487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.660595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.660639 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.660666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.660683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.764085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.764140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.764156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.764180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.764196 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.866799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.866869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.866891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.866920 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.866940 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.905117 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:31 crc kubenswrapper[4744]: E1201 08:18:31.905268 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:31 crc kubenswrapper[4744]: E1201 08:18:31.905334 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:35.905317621 +0000 UTC m=+47.894375552 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.969654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.969726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.969749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.969779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:31 crc kubenswrapper[4744]: I1201 08:18:31.969802 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:31Z","lastTransitionTime":"2025-12-01T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.072186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.072248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.072271 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.072300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.072322 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.174508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.174562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.174578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.174601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.174618 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.277697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.277754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.277771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.277792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.277809 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.284520 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.284609 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.284516 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.284528 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:32 crc kubenswrapper[4744]: E1201 08:18:32.284740 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:32 crc kubenswrapper[4744]: E1201 08:18:32.284852 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:32 crc kubenswrapper[4744]: E1201 08:18:32.284949 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:32 crc kubenswrapper[4744]: E1201 08:18:32.285119 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.380952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.381334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.381551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.381710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.381851 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.484299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.484351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.484369 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.484394 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.484450 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.586827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.586895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.586915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.586944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.586965 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.690628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.690703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.690721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.690747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.690771 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.793928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.794008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.794032 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.794068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.794091 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.897505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.897566 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.897585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.897610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:32 crc kubenswrapper[4744]: I1201 08:18:32.897627 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:32Z","lastTransitionTime":"2025-12-01T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.000735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.000803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.000821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.000851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.000870 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.104576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.104659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.104683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.104714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.104736 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.207948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.208023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.208044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.208074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.208097 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.311204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.311267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.311284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.311307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.311324 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.414588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.414638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.414654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.414676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.414693 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.517114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.517182 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.517207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.517236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.517256 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.620336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.620454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.620480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.620509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.620526 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.724119 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.724167 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.724183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.724206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.724222 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.827364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.827408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.827455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.827479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.827496 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.930736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.930792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.930809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.930833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:33 crc kubenswrapper[4744]: I1201 08:18:33.930852 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:33Z","lastTransitionTime":"2025-12-01T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.033992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.034063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.034078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.034103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.034117 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.138303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.138380 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.138398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.138467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.138485 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.242178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.242246 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.242267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.242301 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.242325 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.284308 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.284305 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.284456 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:34 crc kubenswrapper[4744]: E1201 08:18:34.284494 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.284553 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:34 crc kubenswrapper[4744]: E1201 08:18:34.284732 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:34 crc kubenswrapper[4744]: E1201 08:18:34.284888 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:34 crc kubenswrapper[4744]: E1201 08:18:34.285081 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.344553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.344606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.344623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.344646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.344663 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.448423 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.448843 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.448971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.449126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.449266 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.552391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.552510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.552538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.552571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.552594 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.655888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.655957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.655974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.656000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.656019 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.759607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.759669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.759686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.759713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.759732 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.862482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.863498 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.863710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.863891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.864060 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.967191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.967258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.967275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.967300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:34 crc kubenswrapper[4744]: I1201 08:18:34.967317 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:34Z","lastTransitionTime":"2025-12-01T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.070085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.070142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.070160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.070184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.070200 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.173117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.173598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.173753 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.173977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.174177 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.277698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.277757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.277779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.277810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.277837 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.381143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.381221 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.381238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.381270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.381287 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.484490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.484550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.484567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.484590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.484607 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.587608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.587959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.588148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.588326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.588552 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.691293 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.691369 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.691392 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.691464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.691489 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.794598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.794655 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.794675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.794700 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.794717 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.897963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.898054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.898072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.898096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.898112 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:35Z","lastTransitionTime":"2025-12-01T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:35 crc kubenswrapper[4744]: I1201 08:18:35.956902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:35 crc kubenswrapper[4744]: E1201 08:18:35.957601 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:35 crc kubenswrapper[4744]: E1201 08:18:35.957924 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:18:43.957888349 +0000 UTC m=+55.946946300 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.000872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.000945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.000963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.000989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.001006 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.104150 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.104583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.104776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.104929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.105056 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.207718 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.207771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.207787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.207810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.207827 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.284763 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.284795 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.284819 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:36 crc kubenswrapper[4744]: E1201 08:18:36.284945 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.284997 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:36 crc kubenswrapper[4744]: E1201 08:18:36.285134 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:36 crc kubenswrapper[4744]: E1201 08:18:36.285233 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:36 crc kubenswrapper[4744]: E1201 08:18:36.285356 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.310076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.310343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.310693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.310983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.311285 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.413887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.413958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.413980 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.414010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.414033 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.517294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.517344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.517364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.517393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.517465 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.621059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.621118 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.621141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.621169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.621189 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.723678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.723736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.723755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.723778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.723794 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.827978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.828379 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.828662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.828854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.829018 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.932855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.933776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.934205 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.934595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:36 crc kubenswrapper[4744]: I1201 08:18:36.934906 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:36Z","lastTransitionTime":"2025-12-01T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.038791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.039254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.039513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.039709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.039928 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.143026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.143373 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.143571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.143778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.144019 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.247008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.247079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.247096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.247122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.247140 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.350878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.350947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.350964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.350991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.351007 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.454077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.454153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.454172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.454196 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.454215 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.557382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.557472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.557490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.557516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.557533 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.659698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.660009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.660103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.660191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.660269 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.763016 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.763081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.763099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.763125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.763144 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.866132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.866193 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.866211 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.866236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.866253 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.969793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.969854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.969880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.969906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:37 crc kubenswrapper[4744]: I1201 08:18:37.969923 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:37Z","lastTransitionTime":"2025-12-01T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.073695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.073762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.073780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.073807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.073826 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.080197 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.080388 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:19:10.08036022 +0000 UTC m=+82.069418181 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.176283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.176340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.176356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.176379 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.176395 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.181596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.181854 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.181989 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:19:10.181959103 +0000 UTC m=+82.171017064 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.279675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.279729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.279746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.279774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.279796 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.283285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.283623 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.283492 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.283987 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.284114 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.283796 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.284176 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.284196 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.284229 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.284258 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.284369 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.284356 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.284494 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.284645 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.285127 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.285212 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:19:10.285181071 +0000 UTC m=+82.274239032 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.285349 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.285935 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.286101 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.286332 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:19:10.285648784 +0000 UTC m=+82.274706745 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:18:38 crc kubenswrapper[4744]: E1201 08:18:38.286633 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:19:10.286602082 +0000 UTC m=+82.275660043 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.306558 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.328881 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.346285 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.365136 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.382226 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.382285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.382307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.382336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.382359 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.388843 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.408541 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.429367 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.447298 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.462837 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.484719 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.485187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.485242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.485340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.485374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.485396 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.499508 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.513881 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.531669 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.558224 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.571840 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.588136 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:38Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.588533 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.588571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.588585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.588607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.588623 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.691476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.691542 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.691568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.691596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.691617 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.794556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.794625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.794642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.794667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.794686 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.897513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.897576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.897594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.897618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:38 crc kubenswrapper[4744]: I1201 08:18:38.897636 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:38Z","lastTransitionTime":"2025-12-01T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.000629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.000696 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.000717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.000747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.000769 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.103551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.103595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.103603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.103619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.103628 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.207041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.207113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.207130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.207157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.207175 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.285933 4744 scope.go:117] "RemoveContainer" containerID="52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.310125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.310173 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.310189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.310213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.310230 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.413399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.413522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.413540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.413561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.413576 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.515297 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.515334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.515344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.515360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.515370 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.618050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.618091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.618103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.618120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.618131 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.702932 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/1.log" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.706516 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.706652 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.722313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.722378 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.722395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.722445 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.722462 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.732328 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.758322 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.778561 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.824492 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.824559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.824576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.824971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.825021 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.828884 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.845887 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.864233 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.883380 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.904584 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.916774 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.927812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.927861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.927873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.927895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.927908 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:39Z","lastTransitionTime":"2025-12-01T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.931039 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.948443 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.963576 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.976682 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:39 crc kubenswrapper[4744]: I1201 08:18:39.991823 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.007864 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.024679 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.030401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.030480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.030494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.030512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.030523 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.133390 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.133460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.133475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.133496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.133512 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.140186 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.168447 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.168635 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.191368 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.208404 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.222668 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.231745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.231789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.231802 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.231825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.231838 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.243662 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.248117 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.253279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.253350 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.253372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.253399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.253459 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.258145 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.270999 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.273286 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.281340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.281442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.281460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.281484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.281503 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.284016 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.284048 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.284151 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.284164 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.284195 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.284310 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.284442 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.284545 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.291876 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.299687 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.304390 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.304472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.304490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.304516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.304537 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.309495 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.324004 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.329657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.329724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.329743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.329772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.329789 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.329916 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.342194 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.344989 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.345208 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.347889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.347951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.347975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.348006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.348032 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.361709 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.380800 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.401018 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.418572 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.436265 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.451398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.451482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.451503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.451526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.451546 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.555509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.555551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.555560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.555577 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.555591 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.657972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.658020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.658048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.658072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.658089 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.712475 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/2.log" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.713753 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/1.log" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.718354 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb" exitCode=1 Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.718447 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.718509 4744 scope.go:117] "RemoveContainer" containerID="52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.720750 4744 scope.go:117] "RemoveContainer" containerID="ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb" Dec 01 08:18:40 crc kubenswrapper[4744]: E1201 08:18:40.721193 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.740581 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.762034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.762106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.762122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.762152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.762168 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.764242 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.787296 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.807675 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.825226 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.843647 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.860166 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.870341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.870395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.870449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.870474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.870492 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.884186 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.906041 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.933821 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.953463 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.973663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.973710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.973728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.973752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.973770 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:40Z","lastTransitionTime":"2025-12-01T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.976293 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:40 crc kubenswrapper[4744]: I1201 08:18:40.988002 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.001626 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.013059 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.026702 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.057051 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52dc72985ab6810c5f9bc8f3ee70c7db1638d521af2528967de16f6dcd041c1a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:25Z\\\",\\\"message\\\":\\\"{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1201 08:18:25.568139 6195 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:18:25.568119 6195 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:18:25.568202 6195 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.077557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.077605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.077622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.077687 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.077705 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.179793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.179855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.179872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.179897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.179913 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.283262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.283322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.283343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.283370 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.283388 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.386900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.386960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.386979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.387004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.387021 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.490821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.490886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.490904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.490930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.490947 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.594338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.594385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.594401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.594466 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.594485 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.697666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.697790 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.697817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.697847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.697869 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.724939 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/2.log" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.800729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.800784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.800807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.800840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.800863 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.903559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.903911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.904106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.904283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:41 crc kubenswrapper[4744]: I1201 08:18:41.904502 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:41Z","lastTransitionTime":"2025-12-01T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.008669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.008741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.008760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.008791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.008819 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.112535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.112619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.112640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.112672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.112698 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.216869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.216937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.216956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.216986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.217010 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.284525 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.284578 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:42 crc kubenswrapper[4744]: E1201 08:18:42.284746 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.284773 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:42 crc kubenswrapper[4744]: E1201 08:18:42.284874 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.284942 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:42 crc kubenswrapper[4744]: E1201 08:18:42.285082 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:42 crc kubenswrapper[4744]: E1201 08:18:42.285243 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.320313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.320367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.320383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.320411 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.320458 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.423569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.423625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.423643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.423667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.423686 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.526313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.526375 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.526397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.526462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.526488 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.629478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.629555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.629579 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.629603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.629620 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.732929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.732988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.733004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.733028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.733049 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.836131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.836211 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.836235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.836270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.836291 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.939146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.939229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.939251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.939274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:42 crc kubenswrapper[4744]: I1201 08:18:42.939290 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:42Z","lastTransitionTime":"2025-12-01T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.027946 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.029261 4744 scope.go:117] "RemoveContainer" containerID="ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb" Dec 01 08:18:43 crc kubenswrapper[4744]: E1201 08:18:43.029546 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.042342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.042509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.042532 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.042556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.042576 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.049850 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.067096 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.088523 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.119876 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.135065 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.145744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.145821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.145845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.145877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.145901 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.155879 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.174967 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.194447 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.212087 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.233569 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.249269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.249355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.249380 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.249469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.249499 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.253980 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.272260 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.292545 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.313215 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.331321 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.352829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.352904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.352929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.352958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.352979 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.355645 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.371725 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.455547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.455606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.455632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.455662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.455683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.558283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.558460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.558485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.558553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.558574 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.662393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.662491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.662516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.662550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.662574 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.765526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.765586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.765606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.765628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.765647 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.869212 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.869278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.869296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.869323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.869340 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.972920 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.972994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.973019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.973050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:43 crc kubenswrapper[4744]: I1201 08:18:43.973072 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:43Z","lastTransitionTime":"2025-12-01T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.051334 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:44 crc kubenswrapper[4744]: E1201 08:18:44.051625 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:44 crc kubenswrapper[4744]: E1201 08:18:44.051757 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:19:00.05172645 +0000 UTC m=+72.040784411 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.076172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.076252 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.076282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.076316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.076340 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.179298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.179361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.179379 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.179450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.179477 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.282234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.282294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.282310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.282335 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.282353 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.284649 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.284697 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:44 crc kubenswrapper[4744]: E1201 08:18:44.284826 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.284891 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.284904 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:44 crc kubenswrapper[4744]: E1201 08:18:44.285030 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:44 crc kubenswrapper[4744]: E1201 08:18:44.285210 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:44 crc kubenswrapper[4744]: E1201 08:18:44.285363 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.385719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.385786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.385807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.385838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.385861 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.490152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.490236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.490258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.490288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.490317 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.593593 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.593660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.593682 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.593711 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.593729 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.697582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.697650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.697666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.697693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.697710 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.801133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.801613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.801630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.801652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.801664 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.904676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.904747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.904765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.904791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:44 crc kubenswrapper[4744]: I1201 08:18:44.904813 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:44Z","lastTransitionTime":"2025-12-01T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.007771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.007855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.007881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.007915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.007940 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.111062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.111164 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.111183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.111208 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.111227 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.214114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.214205 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.214240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.214277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.214304 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.317458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.317507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.317517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.317534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.317548 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.420917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.420978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.420990 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.421009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.421022 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.524639 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.524723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.524748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.524780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.524804 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.627620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.627684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.627700 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.627727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.627748 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.731239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.731306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.731322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.731348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.731364 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.834316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.834383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.834449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.834486 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.834510 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.938316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.938391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.938447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.938471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:45 crc kubenswrapper[4744]: I1201 08:18:45.938515 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:45Z","lastTransitionTime":"2025-12-01T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.042334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.042401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.042462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.042487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.042504 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.145450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.145539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.145574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.145608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.145632 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.248479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.248543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.248561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.248587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.248605 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.284095 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.284238 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:46 crc kubenswrapper[4744]: E1201 08:18:46.284450 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.284474 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.284536 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:46 crc kubenswrapper[4744]: E1201 08:18:46.284740 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:46 crc kubenswrapper[4744]: E1201 08:18:46.284807 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:46 crc kubenswrapper[4744]: E1201 08:18:46.284914 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.350800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.350878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.350907 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.350941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.350966 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.454187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.454281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.454305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.454336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.454358 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.557592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.557625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.557634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.557647 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.557657 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.660951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.661027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.661054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.661083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.661104 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.763973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.764092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.764117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.764148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.764172 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.867388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.867513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.867531 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.867559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.867576 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.980838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.981529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.981576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.981601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:46 crc kubenswrapper[4744]: I1201 08:18:46.981619 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:46Z","lastTransitionTime":"2025-12-01T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.083960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.084017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.084029 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.084045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.084054 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.186371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.186399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.186437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.186458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.186470 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.288730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.288787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.288803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.288824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.288841 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.391979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.392042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.392066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.392094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.392117 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.495080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.495134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.495152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.495178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.495195 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.598656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.598721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.598744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.598779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.598803 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.702480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.702559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.702591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.702623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.702647 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.806227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.806289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.806310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.806338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.806352 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.909707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.909782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.909801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.909826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:47 crc kubenswrapper[4744]: I1201 08:18:47.909847 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:47Z","lastTransitionTime":"2025-12-01T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.013038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.013104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.013122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.013148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.013165 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.116240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.116314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.116341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.116371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.116394 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.219906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.219964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.219982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.220005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.220023 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.284453 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.284493 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.285025 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:48 crc kubenswrapper[4744]: E1201 08:18:48.285270 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.285384 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:48 crc kubenswrapper[4744]: E1201 08:18:48.285619 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:48 crc kubenswrapper[4744]: E1201 08:18:48.285784 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:48 crc kubenswrapper[4744]: E1201 08:18:48.285921 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.304862 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.322834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.322911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.322935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.323001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.323023 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.328397 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.348719 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.370074 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.391049 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.406353 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.416801 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.426036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.426066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.426078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.426096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.426106 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.429440 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.440813 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.458735 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.474528 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.489671 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.503248 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.516718 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.528929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.528964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.528974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.528991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.529004 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.529656 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.545103 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.570012 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.631516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.632136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.632223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.632427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.632522 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.735795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.735862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.735880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.735905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.735926 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.860147 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.860216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.860233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.860262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.860282 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.962973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.963037 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.963056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.963082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:48 crc kubenswrapper[4744]: I1201 08:18:48.963099 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:48Z","lastTransitionTime":"2025-12-01T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.066514 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.066585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.066603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.066629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.066649 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.168886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.168962 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.168988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.169019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.169041 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.272341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.272402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.272451 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.272477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.272495 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.375625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.375684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.375702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.375727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.375745 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.485124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.485191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.485210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.485235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.485251 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.587748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.587844 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.587863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.587891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.587909 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.690845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.690905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.690923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.690949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.690966 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.793649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.793698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.793710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.793728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.793743 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.897180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.897246 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.897269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.897300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:49 crc kubenswrapper[4744]: I1201 08:18:49.897328 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:49Z","lastTransitionTime":"2025-12-01T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.000881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.000949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.000970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.000996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.001013 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.103650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.103708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.103721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.103745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.103759 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.207377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.207445 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.207457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.207475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.207486 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.284144 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.284232 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.284174 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.284167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.284369 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.284484 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.284648 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.284747 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.310391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.310498 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.310518 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.310563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.310583 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.413783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.413850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.413874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.413908 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.413932 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.455371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.455478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.455505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.455537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.455560 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.473758 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.479658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.479721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.479746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.479774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.479797 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.497285 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.501730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.501807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.501833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.501864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.501886 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.520215 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.525317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.525446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.525476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.525505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.525527 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.539850 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.543692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.543739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.543756 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.543779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.543799 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.555332 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:50 crc kubenswrapper[4744]: E1201 08:18:50.555611 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.557489 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.557534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.557553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.557577 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.557594 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.659832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.659880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.659896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.659918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.659934 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.768725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.768906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.768920 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.768940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.768958 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.872603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.872660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.872676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.872700 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.872716 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.975578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.975644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.975657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.975674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:50 crc kubenswrapper[4744]: I1201 08:18:50.975686 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:50Z","lastTransitionTime":"2025-12-01T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.078282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.078339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.078352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.078371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.078382 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.180929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.181041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.181064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.181095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.181116 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.283861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.283929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.283954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.283986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.284007 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.385989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.386373 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.386390 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.386427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.386439 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.489160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.489214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.489232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.489256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.489275 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.591992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.592048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.592064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.592089 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.592109 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.694601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.694653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.694670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.694694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.694711 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.797312 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.797375 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.797398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.797464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.797488 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.899244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.899291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.899302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.899318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:51 crc kubenswrapper[4744]: I1201 08:18:51.899326 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:51Z","lastTransitionTime":"2025-12-01T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.002278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.002329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.002347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.002370 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.002389 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.104455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.104501 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.104516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.104539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.104555 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.206551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.206587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.206595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.206610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.206619 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.284732 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.284771 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.284814 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.284845 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:52 crc kubenswrapper[4744]: E1201 08:18:52.284953 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:52 crc kubenswrapper[4744]: E1201 08:18:52.285076 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:52 crc kubenswrapper[4744]: E1201 08:18:52.285176 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:52 crc kubenswrapper[4744]: E1201 08:18:52.285243 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.308297 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.308339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.308356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.308374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.308386 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.411393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.411477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.411488 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.411506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.411517 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.514819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.514880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.514902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.514928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.514947 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.617125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.617197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.617214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.617242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.617262 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.719706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.719759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.719775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.719793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.719806 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.822216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.822256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.822267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.822282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.822292 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.924923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.924986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.925005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.925031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:52 crc kubenswrapper[4744]: I1201 08:18:52.925049 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:52Z","lastTransitionTime":"2025-12-01T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.028105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.028309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.028385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.028486 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.028545 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.131658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.131729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.131749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.131779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.131799 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.235008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.235092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.235111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.235144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.235167 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.338214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.338304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.338324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.338361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.338384 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.442300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.442386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.442432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.442462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.442480 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.545782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.545857 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.545876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.545905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.545924 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.648694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.648727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.648736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.648752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.648760 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.752013 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.752184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.752209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.752281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.752305 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.855181 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.855244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.855258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.855280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.855296 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.959004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.959082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.959103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.959129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:53 crc kubenswrapper[4744]: I1201 08:18:53.959149 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:53Z","lastTransitionTime":"2025-12-01T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.061712 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.061781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.061801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.061846 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.061875 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.165153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.165217 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.165238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.165265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.165284 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.268798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.268860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.268878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.268902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.268919 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.284388 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.284449 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.284521 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:54 crc kubenswrapper[4744]: E1201 08:18:54.284584 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.284605 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:54 crc kubenswrapper[4744]: E1201 08:18:54.284721 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:54 crc kubenswrapper[4744]: E1201 08:18:54.284917 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:54 crc kubenswrapper[4744]: E1201 08:18:54.285156 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.371972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.372052 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.372078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.372110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.372133 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.474809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.474854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.474862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.474877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.474888 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.577986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.578027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.578039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.578056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.578065 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.681186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.681245 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.681264 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.681287 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.681304 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.784080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.784136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.784148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.784166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.784178 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.887144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.887213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.887235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.887265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.887281 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.990897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.990935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.990946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.990962 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:54 crc kubenswrapper[4744]: I1201 08:18:54.990973 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:54Z","lastTransitionTime":"2025-12-01T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.094525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.094599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.094626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.094658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.094681 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.196869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.196953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.196976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.197010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.197030 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.300284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.300355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.300371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.300397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.300440 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.403626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.403678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.403689 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.403707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.403718 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.505567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.505679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.505703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.505726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.505747 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.609252 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.609313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.609331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.609358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.609375 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.712697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.712744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.712755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.712776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.712788 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.816364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.816446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.816462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.816486 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.816502 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.919051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.919098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.919109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.919124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:55 crc kubenswrapper[4744]: I1201 08:18:55.919136 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:55Z","lastTransitionTime":"2025-12-01T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.021857 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.021912 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.021921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.021936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.021946 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.128985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.129041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.129051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.129069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.129078 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.232037 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.232107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.232126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.232151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.232172 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.284668 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:56 crc kubenswrapper[4744]: E1201 08:18:56.284876 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.285133 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:56 crc kubenswrapper[4744]: E1201 08:18:56.285236 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.285523 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:56 crc kubenswrapper[4744]: E1201 08:18:56.285658 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.286012 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:56 crc kubenswrapper[4744]: E1201 08:18:56.286185 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.287633 4744 scope.go:117] "RemoveContainer" containerID="ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb" Dec 01 08:18:56 crc kubenswrapper[4744]: E1201 08:18:56.287990 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.335115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.335174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.335191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.335214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.335232 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.439053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.439111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.439129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.439154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.439171 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.542207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.542238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.542248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.542262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.542271 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.645651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.645749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.645768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.645798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.645816 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.749003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.749063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.749087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.749118 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.749142 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.853055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.853114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.853134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.853159 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.853176 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.956195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.956302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.956328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.956360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:56 crc kubenswrapper[4744]: I1201 08:18:56.956386 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:56Z","lastTransitionTime":"2025-12-01T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.059721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.059842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.059864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.059893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.059910 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.163534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.163589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.163605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.163629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.163645 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.267278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.267336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.267354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.267383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.267431 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.298542 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.370296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.370430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.370449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.370474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.370491 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.474083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.474145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.474162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.474188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.474209 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.576674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.576723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.576736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.576755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.576769 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.679766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.679821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.679830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.679847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.679857 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.782678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.782743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.782753 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.782770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.782780 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.885578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.885712 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.885727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.885744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.885758 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.988849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.989086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.989100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.989114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:57 crc kubenswrapper[4744]: I1201 08:18:57.989124 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:57Z","lastTransitionTime":"2025-12-01T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.091882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.091924 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.091933 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.091950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.091961 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.195743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.195831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.195851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.195873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.195893 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.284647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.284779 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:18:58 crc kubenswrapper[4744]: E1201 08:18:58.284875 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.284917 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.284917 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:18:58 crc kubenswrapper[4744]: E1201 08:18:58.285076 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:18:58 crc kubenswrapper[4744]: E1201 08:18:58.285236 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:18:58 crc kubenswrapper[4744]: E1201 08:18:58.285273 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.298356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.298385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.298392 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.298421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.298431 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.300017 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.315861 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.336161 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.349969 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.363908 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.377588 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.394025 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.400371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.400457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.400480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.400506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.400527 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.411521 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.427742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.443785 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.461819 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.479186 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.492830 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.502839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.502893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.502909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.502934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.502952 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.510165 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.522503 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.536986 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.564900 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.579268 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:18:58Z is after 2025-08-24T17:21:41Z" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.605034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.605119 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.605139 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.605166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.605188 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.707667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.707730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.707749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.707773 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.707791 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.810289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.810347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.810361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.810384 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.810399 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.912803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.912889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.912899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.912923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:58 crc kubenswrapper[4744]: I1201 08:18:58.912938 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:58Z","lastTransitionTime":"2025-12-01T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.015653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.015704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.015713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.015730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.015741 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.118547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.118633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.118657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.118680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.118699 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.220281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.220329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.220338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.220353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.220364 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.322827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.323102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.323183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.323264 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.323362 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.426195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.426251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.426272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.426299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.426319 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.528309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.528443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.528496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.528525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.528545 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.630940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.630978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.630986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.631002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.631011 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.739130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.739167 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.739176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.739190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.739199 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.841122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.841398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.841597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.841759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.841895 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.944133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.944435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.944539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.944652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:18:59 crc kubenswrapper[4744]: I1201 08:18:59.944751 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:18:59Z","lastTransitionTime":"2025-12-01T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.100336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.100375 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.100386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.100401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.100426 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.143937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.144096 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.144161 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:19:32.144144665 +0000 UTC m=+104.133202586 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.203069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.203359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.203475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.203571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.203638 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.284344 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.284344 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.284506 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.284751 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.284801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.284986 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.285115 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.285182 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.305868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.305901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.305911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.305926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.305936 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.408759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.408818 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.408837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.408862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.408881 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.511059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.511103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.511113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.511131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.511140 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.614353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.614437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.614456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.614479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.614496 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.617445 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.617506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.617525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.617549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.617568 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.632337 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.637889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.637930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.637939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.637954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.637964 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.659400 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.663578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.663632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.663652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.663679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.663697 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.684090 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.688907 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.688945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.688961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.688984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.689001 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.707010 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.710034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.710066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.710075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.710088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.710097 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.722952 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:00 crc kubenswrapper[4744]: E1201 08:19:00.723102 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.724450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.724482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.724493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.724509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.724520 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.826346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.826399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.826443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.826468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.826485 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.929388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.929467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.929485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.929510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:00 crc kubenswrapper[4744]: I1201 08:19:00.929529 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:00Z","lastTransitionTime":"2025-12-01T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.032965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.033023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.033042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.033064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.033080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.137311 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.137355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.137365 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.137381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.137392 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.240086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.240149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.240162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.240192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.240208 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.342592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.342643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.342655 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.342675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.342691 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.446474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.446560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.446581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.446609 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.446678 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.549404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.549506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.549525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.549552 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.549570 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.652634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.652702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.652718 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.652743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.652761 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.755625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.755697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.755714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.755742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.755758 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.804317 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/0.log" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.804387 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a5e5532-8675-4f3f-aa30-7c22bc95b311" containerID="49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9" exitCode=1 Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.804544 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerDied","Data":"49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.805069 4744 scope.go:117] "RemoveContainer" containerID="49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.823166 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.840956 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.859142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.859186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.859204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.859228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.859247 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.859350 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.879768 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.903215 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.920595 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.942333 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.962969 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.963334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.963362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.963378 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.963400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.963449 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:01Z","lastTransitionTime":"2025-12-01T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:01 crc kubenswrapper[4744]: I1201 08:19:01.991050 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:01Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.006352 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.025401 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.054940 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.065836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.065889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.065907 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.065932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.065949 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.070149 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.091783 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.112517 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.130766 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.147794 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.164087 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.168505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.168563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.168581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.168604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.168622 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.271951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.272005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.272018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.272041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.272057 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.284719 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.284775 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.284746 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:02 crc kubenswrapper[4744]: E1201 08:19:02.284914 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.285014 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:02 crc kubenswrapper[4744]: E1201 08:19:02.285144 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:02 crc kubenswrapper[4744]: E1201 08:19:02.285322 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:02 crc kubenswrapper[4744]: E1201 08:19:02.285474 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.374634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.374705 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.374723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.374750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.374772 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.477637 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.477695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.477713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.477740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.477756 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.580892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.580951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.580967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.580992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.581011 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.683723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.683792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.683815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.683847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.683866 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.786512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.786585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.786604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.786628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.786647 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.811227 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/0.log" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.811313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerStarted","Data":"e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.831769 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.856873 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.878501 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.889641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.889710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.889729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.889754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.889772 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.899849 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.928742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.945950 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.966155 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.984998 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:02Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.993146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.993200 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.993223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.993254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:02 crc kubenswrapper[4744]: I1201 08:19:02.993276 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:02Z","lastTransitionTime":"2025-12-01T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.003369 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.020747 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.037478 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.056991 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.073456 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.094460 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.096877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.096925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.096943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.096966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.096984 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.109783 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.130953 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.148568 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.171631 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:03Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.200156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.200203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.200219 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.200241 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.200258 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.302765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.302804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.302813 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.302829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.302839 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.405859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.405892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.405900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.405914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.405923 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.507870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.507940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.507957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.507982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.508003 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.611127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.611190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.611207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.611234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.611254 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.714674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.715097 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.715294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.715569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.715770 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.818481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.818529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.818571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.818596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.818612 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.963813 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.963874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.963894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.963931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:03 crc kubenswrapper[4744]: I1201 08:19:03.963950 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:03Z","lastTransitionTime":"2025-12-01T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.066762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.066838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.066858 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.066883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.066902 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.169562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.169656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.169674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.169698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.169714 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.272651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.272717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.272740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.272771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.272792 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.284936 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.284988 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:04 crc kubenswrapper[4744]: E1201 08:19:04.285108 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.284947 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:04 crc kubenswrapper[4744]: E1201 08:19:04.285265 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:04 crc kubenswrapper[4744]: E1201 08:19:04.285395 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.285600 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:04 crc kubenswrapper[4744]: E1201 08:19:04.285831 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.375026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.375073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.375090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.375112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.375129 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.479117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.479169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.479186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.479209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.479227 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.582062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.582128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.582164 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.582193 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.582217 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.686062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.686156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.686189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.686234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.686276 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.790022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.790412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.790667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.790881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.791104 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.894795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.894869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.894893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.894923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.894943 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.998804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.999558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.999610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.999643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:04 crc kubenswrapper[4744]: I1201 08:19:04.999670 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:04Z","lastTransitionTime":"2025-12-01T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.103236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.103309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.103326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.103353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.103373 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.206090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.206162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.206187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.206216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.206239 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.310129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.310204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.310222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.310249 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.310268 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.413733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.413816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.413835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.413866 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.413886 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.516736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.516812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.516830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.516860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.516883 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.620352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.620448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.620467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.620493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.620514 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.724339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.724398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.724446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.724480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.724499 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.827289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.827352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.827372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.827392 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.827408 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.930845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.930905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.930922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.930948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:05 crc kubenswrapper[4744]: I1201 08:19:05.930965 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:05Z","lastTransitionTime":"2025-12-01T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.034069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.034137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.034155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.034181 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.034199 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.137157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.137241 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.137263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.137289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.137309 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.239988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.240047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.240064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.240090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.240108 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.284926 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.285054 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.284938 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.285173 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:06 crc kubenswrapper[4744]: E1201 08:19:06.285150 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:06 crc kubenswrapper[4744]: E1201 08:19:06.285264 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:06 crc kubenswrapper[4744]: E1201 08:19:06.285485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:06 crc kubenswrapper[4744]: E1201 08:19:06.285647 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.343513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.343573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.343597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.343626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.343647 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.446535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.446610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.446635 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.446661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.446678 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.549671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.549749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.549769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.549798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.549814 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.652658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.652734 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.652753 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.652778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.652795 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.755798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.755875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.755897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.755925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.755943 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.858881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.858960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.858981 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.859008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.859027 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.961668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.961738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.961764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.961794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:06 crc kubenswrapper[4744]: I1201 08:19:06.961817 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:06Z","lastTransitionTime":"2025-12-01T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.064693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.064752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.064774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.064802 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.064822 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.167388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.167478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.167501 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.167529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.167550 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.272202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.272307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.272331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.272360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.272381 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.376514 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.376584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.376600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.376644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.376663 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.480023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.480093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.480110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.480139 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.480156 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.584292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.584356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.584375 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.584401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.584465 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.688077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.688143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.688160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.688185 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.688204 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.792257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.792317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.792333 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.792357 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.792376 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.896400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.896491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.896508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.896533 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.896551 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.999116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.999176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.999192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.999215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:07 crc kubenswrapper[4744]: I1201 08:19:07.999234 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:07Z","lastTransitionTime":"2025-12-01T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.102335 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.102396 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.102489 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.102556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.102577 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.205837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.205901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.205919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.205946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.205966 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.285136 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.285154 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:08 crc kubenswrapper[4744]: E1201 08:19:08.285297 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.285567 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:08 crc kubenswrapper[4744]: E1201 08:19:08.285670 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:08 crc kubenswrapper[4744]: E1201 08:19:08.285540 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.285776 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:08 crc kubenswrapper[4744]: E1201 08:19:08.285859 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.305536 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.308564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.308617 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.308635 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.308661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.308715 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.324711 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.343388 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.360218 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.377266 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.397641 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.411314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.411353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.411365 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.411381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.411394 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.417842 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.435333 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.450068 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.472285 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.490461 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.514487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.514248 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.514540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.514750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.514778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.514797 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.534337 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.555487 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.572102 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.593308 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.617497 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.617568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.617587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.617611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.617628 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.625170 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.641135 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:08Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.723341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.723399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.723421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.723468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.723487 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.826586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.826628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.826645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.826667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.826683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.929134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.929193 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.929210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.929234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:08 crc kubenswrapper[4744]: I1201 08:19:08.929252 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:08Z","lastTransitionTime":"2025-12-01T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.032068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.032132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.032151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.032176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.032194 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.135698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.135749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.135771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.135798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.135815 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.238497 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.238553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.238566 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.238587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.238601 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.341795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.341864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.341882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.341906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.341924 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.445155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.445255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.445273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.445334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.445353 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.549057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.549120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.549134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.549155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.549170 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.652630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.652707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.652730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.652763 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.652788 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.756378 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.756487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.756508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.756533 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.756550 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.859938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.860006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.860023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.860048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.860066 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.963182 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.963259 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.963286 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.963309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:09 crc kubenswrapper[4744]: I1201 08:19:09.963324 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:09Z","lastTransitionTime":"2025-12-01T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.066562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.066625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.066642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.066668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.066685 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.158210 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.158503 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:14.158465643 +0000 UTC m=+146.147523594 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.170559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.170660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.170682 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.170737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.170755 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.259561 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.259689 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.259767 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:20:14.259747281 +0000 UTC m=+146.248805232 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.274577 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.274679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.274728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.274752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.274768 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.284144 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.284167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.284496 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.284738 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.285627 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.285853 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.286254 4744 scope.go:117] "RemoveContainer" containerID="ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.286839 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.287473 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.360448 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.360664 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.360678 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.360753 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.360780 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:20:14.360753471 +0000 UTC m=+146.349811432 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.360931 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.360963 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.360985 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.361057 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:20:14.361034449 +0000 UTC m=+146.350092410 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.361065 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.361088 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.361101 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.361141 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:20:14.361127072 +0000 UTC m=+146.350184993 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.377438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.377993 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.378023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.378056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.378080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.480992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.481033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.481046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.481065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.481081 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.583886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.583939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.583970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.583995 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.584041 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.688712 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.688785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.688804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.688827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.688853 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.752210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.752261 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.752272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.752290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.752302 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.774032 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.780669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.780745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.780772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.780804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.780828 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.796544 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.804002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.804045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.804057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.804075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.804089 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.818264 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.821901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.821954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.821969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.821989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.822004 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.833238 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.837822 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.837880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.837893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.837917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.837933 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.841936 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/2.log" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.844986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.845490 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.858165 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: E1201 08:19:10.858317 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.860022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.860053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.860065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.860085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.860098 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.865546 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.882195 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.895154 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.913820 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.928239 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.939859 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.955778 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.962515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.962563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.962577 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.962596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.962610 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:10Z","lastTransitionTime":"2025-12-01T08:19:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.965133 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.977264 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.985593 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:10 crc kubenswrapper[4744]: I1201 08:19:10.997520 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.016771 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.026257 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.036439 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.049838 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.059256 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.065030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.065073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.065085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.065102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.065113 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.068865 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.082168 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.167529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.167583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.167598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.167619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.167633 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.271201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.271263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.271281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.271304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.271323 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.373668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.373713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.373724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.373740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.373749 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.476707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.476770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.476792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.476866 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.476889 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.583935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.584003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.584025 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.584053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.584074 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.687470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.687530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.687549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.687577 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.687595 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.790176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.790242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.790259 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.790284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.790302 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.852168 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/3.log" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.853277 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/2.log" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.857213 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1" exitCode=1 Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.857278 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.857449 4744 scope.go:117] "RemoveContainer" containerID="ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.859633 4744 scope.go:117] "RemoveContainer" containerID="7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1" Dec 01 08:19:11 crc kubenswrapper[4744]: E1201 08:19:11.859901 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.879248 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.893555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.893619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.893641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.893673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.893694 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.894831 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.915582 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.948528 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3659b123d561aa43d9ddbe4710a80f968d0c73da7bdb14b243807284b9baeb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"message\\\":\\\"enshift-ovn-kubernetes/ovnkube-node-2sflk\\\\nI1201 08:18:40.240510 6392 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:11Z\\\",\\\"message\\\":\\\"t:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165168 6788 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:19:11.165283 6788 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165020 6788 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator]} name:Service_openshift-machine-api/machine-api-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:19:11.165522 6788 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:19:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.964849 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.985518 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:11Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.996120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.996191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.996212 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.996239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:11 crc kubenswrapper[4744]: I1201 08:19:11.996259 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:11Z","lastTransitionTime":"2025-12-01T08:19:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.005352 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.025102 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.042782 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.060363 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.076916 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.096182 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.099919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.100011 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.100036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.100072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.100096 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.115133 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.135698 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.159562 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.179901 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.203722 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.204178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.204208 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.204223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.204247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.204265 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.220851 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.284759 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.284847 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.284875 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.284804 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:12 crc kubenswrapper[4744]: E1201 08:19:12.285032 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:12 crc kubenswrapper[4744]: E1201 08:19:12.285125 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:12 crc kubenswrapper[4744]: E1201 08:19:12.285255 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:12 crc kubenswrapper[4744]: E1201 08:19:12.285436 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.308061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.308103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.308119 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.308139 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.308155 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.411094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.411156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.411174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.411197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.411215 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.514227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.514291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.514303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.514318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.514329 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.617829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.617879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.617895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.617920 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.617968 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.720989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.721050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.721067 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.721090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.721109 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.825851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.825923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.825946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.825978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.826002 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.863347 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/3.log" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.868903 4744 scope.go:117] "RemoveContainer" containerID="7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1" Dec 01 08:19:12 crc kubenswrapper[4744]: E1201 08:19:12.869165 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.888448 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.907774 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.926677 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.930482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.930539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.930558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.930582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.930623 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:12Z","lastTransitionTime":"2025-12-01T08:19:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.944937 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.966651 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:12 crc kubenswrapper[4744]: I1201 08:19:12.985870 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:12Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.005942 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.026200 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.033300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.033388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.033440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.033474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.033493 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.040889 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.061565 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.079947 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.104003 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.120910 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.136552 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.136611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.136630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.136654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.136670 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.137366 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.157732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.182302 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:11Z\\\",\\\"message\\\":\\\"t:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165168 6788 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:19:11.165283 6788 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165020 6788 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator]} name:Service_openshift-machine-api/machine-api-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:19:11.165522 6788 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:19:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.199818 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.223705 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:13Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.239048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.239113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.239130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.239154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.239172 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.341887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.341976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.341997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.342026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.342047 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.444590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.444644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.444659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.444684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.444704 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.547848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.547955 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.547981 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.548011 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.548034 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.650583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.650653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.650671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.650696 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.650717 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.754204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.754571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.754594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.754627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.754647 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.857965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.858024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.858040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.858068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.858086 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.961126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.961190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.961214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.961248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:13 crc kubenswrapper[4744]: I1201 08:19:13.961271 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:13Z","lastTransitionTime":"2025-12-01T08:19:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.063998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.064056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.064075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.064100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.064117 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.167143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.167199 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.167216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.167239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.167257 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.271108 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.271175 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.271194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.271219 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.271238 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.284770 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.284833 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.284942 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:14 crc kubenswrapper[4744]: E1201 08:19:14.285105 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.285198 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:14 crc kubenswrapper[4744]: E1201 08:19:14.285246 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:14 crc kubenswrapper[4744]: E1201 08:19:14.285472 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:14 crc kubenswrapper[4744]: E1201 08:19:14.285586 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.374640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.374719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.374745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.374777 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.374801 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.478216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.478299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.478324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.478355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.478381 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.581075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.581131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.581147 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.581172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.581188 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.684362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.684535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.684555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.684580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.684604 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.788256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.788326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.788343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.788367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.788385 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.891303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.891385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.891491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.891561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.891582 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.995670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.995736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.995752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.995778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:14 crc kubenswrapper[4744]: I1201 08:19:14.995796 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:14Z","lastTransitionTime":"2025-12-01T08:19:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.099196 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.099258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.099275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.099299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.099317 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.201510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.201586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.201597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.201613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.201624 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.304308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.304343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.304352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.304364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.304374 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.406827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.406868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.406877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.406891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.406899 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.509751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.509810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.509870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.509896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.509917 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.612524 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.612590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.612606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.612630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.612648 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.715985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.716043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.716060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.716082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.716101 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.818906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.818966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.818983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.819010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.819027 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.921873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.921937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.921955 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.921982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:15 crc kubenswrapper[4744]: I1201 08:19:15.922000 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:15Z","lastTransitionTime":"2025-12-01T08:19:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.025138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.025210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.025232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.025263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.025282 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.128638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.128715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.128732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.128755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.128771 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.232203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.232265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.232282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.232308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.232326 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.285001 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.285100 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.285108 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.285194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:16 crc kubenswrapper[4744]: E1201 08:19:16.285262 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:16 crc kubenswrapper[4744]: E1201 08:19:16.285581 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:16 crc kubenswrapper[4744]: E1201 08:19:16.285714 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:16 crc kubenswrapper[4744]: E1201 08:19:16.285873 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.334403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.334491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.334509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.334532 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.334550 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.437897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.437976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.437999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.438029 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.438056 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.542482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.542540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.542557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.542580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.542597 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.646479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.646551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.646573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.646603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.646623 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.749464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.749521 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.749534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.749554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.749567 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.852500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.852572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.852584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.852600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.852912 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.955872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.955927 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.955948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.955976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:16 crc kubenswrapper[4744]: I1201 08:19:16.955998 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:16Z","lastTransitionTime":"2025-12-01T08:19:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.058879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.058948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.058965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.058992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.059010 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.162010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.162074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.162097 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.162121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.162141 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.265393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.265499 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.265520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.265546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.265564 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.368310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.368381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.368395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.368436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.368451 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.471764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.471838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.471864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.471895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.471918 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.574769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.574830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.574852 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.574881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.574903 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.677667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.678083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.678331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.678560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.678745 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.782530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.782738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.782796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.782824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.782842 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.885504 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.885563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.885581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.885611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.885630 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.988937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.989003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.989020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.989046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:17 crc kubenswrapper[4744]: I1201 08:19:17.989064 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:17Z","lastTransitionTime":"2025-12-01T08:19:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.091744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.091810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.091821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.091840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.091851 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.194534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.194616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.194639 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.194677 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.194699 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.284602 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.284617 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.284767 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:18 crc kubenswrapper[4744]: E1201 08:19:18.284942 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:18 crc kubenswrapper[4744]: E1201 08:19:18.285121 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.285653 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:18 crc kubenswrapper[4744]: E1201 08:19:18.286066 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:18 crc kubenswrapper[4744]: E1201 08:19:18.286472 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.297285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.297331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.297343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.297361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.297373 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.307369 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.308347 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.326133 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.349742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.392552 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:11Z\\\",\\\"message\\\":\\\"t:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165168 6788 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:19:11.165283 6788 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165020 6788 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator]} name:Service_openshift-machine-api/machine-api-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:19:11.165522 6788 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:19:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.399367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.399427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.399437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.399455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.399466 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.419700 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.446876 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.462324 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.477701 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.490014 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.499766 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.501778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.501828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.501839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.501860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.501885 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.511452 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.523336 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.536157 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.548729 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.562838 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.575733 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.592921 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.604294 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:18Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.604814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.604840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.604851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.604871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.604914 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.707682 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.707753 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.707768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.707785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.707798 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.810607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.811074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.811248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.811467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.811670 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.915798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.916180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.916346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.916546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:18 crc kubenswrapper[4744]: I1201 08:19:18.916687 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:18Z","lastTransitionTime":"2025-12-01T08:19:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.020290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.020370 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.020389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.020450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.020471 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.123666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.123735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.123752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.123784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.123803 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.227899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.228290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.228309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.228335 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.228357 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.330889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.330960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.330977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.331003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.331020 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.434188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.434243 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.434261 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.434285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.434304 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.537626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.537691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.537713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.537740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.537762 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.641038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.641090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.641106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.641128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.641154 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.744485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.744549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.744565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.744588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.744607 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.847774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.847838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.847855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.847885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.847904 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.951068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.951161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.951179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.951207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:19 crc kubenswrapper[4744]: I1201 08:19:19.951226 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:19Z","lastTransitionTime":"2025-12-01T08:19:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.054672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.054723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.054743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.054764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.054780 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.158076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.158154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.158173 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.158199 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.158217 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.261934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.261995 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.262012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.262036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.262089 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.285069 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.285158 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.285091 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:20 crc kubenswrapper[4744]: E1201 08:19:20.285247 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:20 crc kubenswrapper[4744]: E1201 08:19:20.285354 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:20 crc kubenswrapper[4744]: E1201 08:19:20.285493 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.286071 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:20 crc kubenswrapper[4744]: E1201 08:19:20.286391 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.365339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.365391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.365433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.365458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.365477 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.468702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.468755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.468772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.468791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.468807 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.571620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.571678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.571693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.571717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.571739 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.674805 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.674886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.674911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.674938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.674959 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.778715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.778796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.778823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.778854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.778879 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.881975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.882042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.882060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.882084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.882103 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.985453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.985496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.985507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.985524 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:20 crc kubenswrapper[4744]: I1201 08:19:20.985535 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:20Z","lastTransitionTime":"2025-12-01T08:19:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.040137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.040216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.040242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.040274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.040295 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: E1201 08:19:21.065283 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.070687 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.070796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.070817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.070888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.070908 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: E1201 08:19:21.092824 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.097840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.097909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.097928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.097958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.097977 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: E1201 08:19:21.118701 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.124086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.124149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.124166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.124192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.124210 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: E1201 08:19:21.145095 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.150786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.150848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.150865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.150889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.150907 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: E1201 08:19:21.171260 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:21Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:21 crc kubenswrapper[4744]: E1201 08:19:21.171555 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.173796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.173855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.173873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.173899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.173918 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.277270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.277338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.277360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.277387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.277436 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.380433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.380503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.380520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.380545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.380563 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.483780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.483846 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.483864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.483891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.483909 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.587019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.587077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.587093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.587116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.587135 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.690132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.690184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.690207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.690236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.690257 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.793841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.793903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.793922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.793947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.793963 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.896880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.896966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.896992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.897022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.897043 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.999372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.999457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.999484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.999509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:21 crc kubenswrapper[4744]: I1201 08:19:21.999531 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:21Z","lastTransitionTime":"2025-12-01T08:19:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.102734 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.102793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.102811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.102836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.102853 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.205811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.205863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.205879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.205903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.205919 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.284918 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.284991 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:22 crc kubenswrapper[4744]: E1201 08:19:22.285097 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.285130 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.284916 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:22 crc kubenswrapper[4744]: E1201 08:19:22.285217 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:22 crc kubenswrapper[4744]: E1201 08:19:22.285325 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:22 crc kubenswrapper[4744]: E1201 08:19:22.285538 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.308843 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.308893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.308913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.308935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.308953 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.412104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.412180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.412198 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.412224 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.412242 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.515526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.515586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.515610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.515642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.515661 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.618739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.618804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.618821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.618848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.618874 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.723652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.723715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.723733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.723756 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.723774 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.827068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.827120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.827138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.827163 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.827180 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.930248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.930308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.930326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.930352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:22 crc kubenswrapper[4744]: I1201 08:19:22.930373 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:22Z","lastTransitionTime":"2025-12-01T08:19:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.033879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.033944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.033962 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.033988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.034009 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.137444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.137526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.137540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.137576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.137594 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.241618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.241699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.241718 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.241745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.241766 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.344960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.345027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.345047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.345074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.345091 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.447613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.447680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.447698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.447726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.447746 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.551324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.551394 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.551441 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.551470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.551488 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.655011 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.655080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.655096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.655122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.655140 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.757964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.758030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.758047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.758072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.758094 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.861175 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.861244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.861260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.861287 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.861307 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.964494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.964572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.964589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.964614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:23 crc kubenswrapper[4744]: I1201 08:19:23.964633 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:23Z","lastTransitionTime":"2025-12-01T08:19:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.068303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.068363 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.068381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.068432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.068451 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.172158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.172267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.172287 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.172316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.172336 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.276071 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.276164 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.276183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.276211 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.276229 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.284589 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.284614 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.284728 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.284779 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:24 crc kubenswrapper[4744]: E1201 08:19:24.284809 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:24 crc kubenswrapper[4744]: E1201 08:19:24.284931 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:24 crc kubenswrapper[4744]: E1201 08:19:24.285040 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:24 crc kubenswrapper[4744]: E1201 08:19:24.285335 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.379631 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.379697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.379714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.379742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.379760 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.482759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.482827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.482845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.482871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.482889 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.586520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.586578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.586597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.586623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.586641 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.689998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.690070 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.690088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.690116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.690136 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.793046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.793136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.793155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.793179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.793197 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.896651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.896727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.896750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.896787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:24 crc kubenswrapper[4744]: I1201 08:19:24.896811 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:24Z","lastTransitionTime":"2025-12-01T08:19:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.000774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.000837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.000858 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.000885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.000902 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.103896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.103960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.103977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.104002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.104020 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.207371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.207461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.207480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.207508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.207528 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.310748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.310827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.310851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.310877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.310896 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.413714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.413784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.413804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.413831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.413849 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.517494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.517547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.517568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.517813 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.517844 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.620901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.620964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.620984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.621014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.621037 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.724195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.724249 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.724267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.724291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.724308 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.827656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.827732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.827759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.827790 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.827814 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.930540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.930594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.930611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.930634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:25 crc kubenswrapper[4744]: I1201 08:19:25.930652 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:25Z","lastTransitionTime":"2025-12-01T08:19:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.038730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.038815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.038840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.038881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.038903 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.142792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.142848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.142867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.142892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.142912 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.245778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.246192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.246609 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.247126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.247467 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.284670 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.284832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.284993 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:26 crc kubenswrapper[4744]: E1201 08:19:26.285013 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.285233 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:26 crc kubenswrapper[4744]: E1201 08:19:26.286085 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:26 crc kubenswrapper[4744]: E1201 08:19:26.286171 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:26 crc kubenswrapper[4744]: E1201 08:19:26.286255 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.286642 4744 scope.go:117] "RemoveContainer" containerID="7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1" Dec 01 08:19:26 crc kubenswrapper[4744]: E1201 08:19:26.286916 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.350841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.350891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.350911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.350936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.350956 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.453941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.454348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.454542 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.454707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.454848 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.558614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.559050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.559228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.559660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.559855 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.662823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.663606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.663644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.663672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.663691 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.766112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.766145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.766156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.766174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.766184 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.869508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.869557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.869567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.869588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.869599 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.972220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.972293 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.972315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.972343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:26 crc kubenswrapper[4744]: I1201 08:19:26.972363 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:26Z","lastTransitionTime":"2025-12-01T08:19:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.075860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.075921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.075939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.075964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.075982 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.179396 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.179503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.179578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.179604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.179661 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.283192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.283260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.283281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.283305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.283322 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.386651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.386743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.386763 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.386788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.386807 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.489220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.489280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.489297 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.489322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.489339 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.592934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.592994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.593013 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.593036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.593055 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.696072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.696145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.696170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.696200 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.696222 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.798937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.799038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.799057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.799080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.799096 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.905941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.906018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.906042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.906069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:27 crc kubenswrapper[4744]: I1201 08:19:27.906088 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:27Z","lastTransitionTime":"2025-12-01T08:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.011225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.011288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.011305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.011331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.011348 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.114286 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.114328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.114339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.114355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.114368 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.217123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.217458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.217631 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.217789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.217918 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.284692 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:28 crc kubenswrapper[4744]: E1201 08:19:28.284875 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.284891 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.284925 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:28 crc kubenswrapper[4744]: E1201 08:19:28.285012 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:28 crc kubenswrapper[4744]: E1201 08:19:28.285131 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.285576 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:28 crc kubenswrapper[4744]: E1201 08:19:28.285779 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.309192 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ceb4de2-f0af-4141-a088-28f6a7525832\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 08:18:00.661087 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:18:00.662950 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472325955/tls.crt::/tmp/serving-cert-3472325955/tls.key\\\\\\\"\\\\nI1201 08:18:06.176137 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 08:18:06.180629 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 08:18:06.180668 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 08:18:06.180704 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 08:18:06.180747 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 08:18:06.193631 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 08:18:06.193672 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193682 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 08:18:06.193690 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 08:18:06.193697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 08:18:06.193703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 08:18:06.193708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 08:18:06.193758 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 08:18:06.197270 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.321744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.321799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.321817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.321841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.321858 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.328361 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://581021aaf3998c6517379a39eca9425042412d9410cb71d3c72a24b8125307fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.352482 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9dd3b9-b551-4a4b-8c4f-12ff6309034c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0acb239cbe21e626592972981405066d0d70f4d8721393e695587c434a12f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f003ad9afc43c23b61d1ab02d0f580cc5204d9e7236bd10e34239c30142caaa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b23e9b5c4758af91094823e371d447f98a24cf516397d2f90556374c8d0e075\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1e22ef75f136359e8aaae5bca65a08f789ddaf15ec32e7242ddde8faebaa0d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7fde57808547cfde2ab630153cc3db188d01d13a0ef2a5f65a24355576c3d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25721ea0fa2f900c5e6a04d8026d9b74c3dccc76116c0b79c631a8912ef71382\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbbefc81122f638a790fe9743411af65173b03e533769e4b464874b0bc8ec5fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qsm67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cvbp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.369606 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p489k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ac81c1-4bf7-472c-8110-21364dde7fd6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfnrb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p489k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.404872 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d981c10-0e30-465f-8a4c-1511195029a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264a09da8207850acee482e1cfc3a509b5b85763227abf3a41512235545913e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eac91565f28c77e5ce801eceae446eb198e48a68054016d5a45f9869fc57171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c75976377e25c4dad812bf4645d19067e30a105e6f7d441f22cbe71f1bc8c6ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a16c0b2005868f2e7421bbf46c59cfb85bb00f878acaf39036f32f7d4eb5bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb17b2f336bc4bff25a29bc95b8d11d380eaa6dc9984588aae31871d61e8928f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e52d273edf1b7384ea71d7252be10669e628bb8e150d32e79f97898133c72a3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e52d273edf1b7384ea71d7252be10669e628bb8e150d32e79f97898133c72a3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5068ca067105692d5312f839b3a530b1bfb5f4908f41ede747b9b5026b38b2a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5068ca067105692d5312f839b3a530b1bfb5f4908f41ede747b9b5026b38b2a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e4d6a48cdf12f626971728b6129cdfdcf7ac240fd2d57f1bbd0769150bf46187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4d6a48cdf12f626971728b6129cdfdcf7ac240fd2d57f1bbd0769150bf46187\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.422846 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncr54" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"199cd5a5-cb9c-4e04-b7fb-da877b86eafc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d519dfbbf824a071d81a5243bbec92509f1a28207876ef27dc0935b147a99dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zmm4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncr54\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.424017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.424073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.424091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.424114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.424131 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.443517 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vsh4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a5e5532-8675-4f3f-aa30-7c22bc95b311\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:01Z\\\",\\\"message\\\":\\\"2025-12-01T08:18:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1\\\\n2025-12-01T08:18:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2416f90-33cb-431e-9381-b47ab55a9da1 to /host/opt/cni/bin/\\\\n2025-12-01T08:18:16Z [verbose] multus-daemon started\\\\n2025-12-01T08:18:16Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:19:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:19:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7nxrc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vsh4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.473316 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1df124eb-0511-4702-85d8-3e324c59bb0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:19:11Z\\\",\\\"message\\\":\\\"t:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165168 6788 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:19:11.165283 6788 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:19:11.165020 6788 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator]} name:Service_openshift-machine-api/machine-api-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 08:19:11.165522 6788 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:19:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxzmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2sflk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.489063 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z4dh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cac4540b-1bb4-43e1-b373-c0bff223afdd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://291224ebe2e90363533e2217026bb9aa201f9d932f0d3c659f6bdb9c9482041c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmlxw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z4dh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.508233 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.526250 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.527512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.527564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.527581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.527607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.527672 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.541493 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.557935 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc42676bf96c841a6eacce8d2ae0d399b617e715201587870704715aa4a3eb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh7zk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g75sq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.575529 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"707e21d6-1ab8-40f2-8681-52a0263cddff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0be01f13f66c0c1ab013e4d12f747fd61ca6e5f556cd7eaf078e6af878606137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3849e6df95694855d4e121feff2fa6fc9c2b19e6ccbbbecf5e4be03e2018a6af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s9d6j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:18:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lwkvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.594629 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://774f5b3d51894c013cc70903830e6134709e506156d59bf8bca1536feccdeb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.612838 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dba81dd-735e-4890-b784-e353a09e5a19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://918b033d42f9c7e2c69c38531ba5b2e752e65358630f26cf7a602ac6f1744e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4911009853e944771cd13bbc9df84e4f46a0b5e06a5f48a426af6f38ef87d56b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e83fd4e4bb4f0d2b816af2f3c9ce736c3b8067e9d771367ac95e03b9c99882e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.628387 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1894e862-1bbf-402a-a299-a3b21a6ec92d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d5d1442ec24655d0d81dc48ff5433ae841a0a10a837356dfa4ff534a6867c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://785cfbd0f8881bb43f0c6f295dba943e62710db13d5700d0da5e6d2f23521acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c9db20fe2619e6bf85e074347f21cbf1f4b461dd6f4c739decf4a8b5aad5efd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdbed0e58142a98bc6f2a8c25e112b091d313d719c703ae43d82aac2724cdab7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.630484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.630539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.630557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.630588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.630604 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.645526 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:18:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d1801e06fee825ecfc6500ec3b888e9b558815de2a8a3dee15d56c36bc8daab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee03f6375cf9a618d8f5f468f0fdd5298d2f8f8277facf2df5011d9fd380077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:18:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.659238 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18a2282b-4c17-4900-998c-fc8c52b6139c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d81fc0b36d6b7f9e9d403e8d655916c8461aba6504586a78bdcd5cdd860334bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://457296a02511c1b4a6a6167fdd5fb0e3fa5e72ccd20fad49bc3e89de4cdf130e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:17:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:17:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:28Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.734027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.734180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.734202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.734231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.734249 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.837519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.837581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.837600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.837625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.837643 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.939675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.939736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.939752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.939776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:28 crc kubenswrapper[4744]: I1201 08:19:28.939794 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:28Z","lastTransitionTime":"2025-12-01T08:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.042979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.043055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.043072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.043096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.043114 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.146714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.146824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.146903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.146937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.146989 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.249851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.249935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.249956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.249981 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.250000 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.352606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.352656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.352672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.352694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.352711 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.454979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.455032 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.455048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.455070 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.455085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.557437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.557499 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.557517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.557542 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.557562 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.660645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.660722 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.660738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.660765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.660786 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.767657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.767774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.767812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.767879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.767898 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.870878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.871255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.871403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.871614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.871780 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.974783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.974849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.974867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.974892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:29 crc kubenswrapper[4744]: I1201 08:19:29.974911 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:29Z","lastTransitionTime":"2025-12-01T08:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.077975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.078037 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.078053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.078076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.078093 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.180832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.180924 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.180942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.181003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.181022 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.284333 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.284439 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.284494 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.284340 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:30 crc kubenswrapper[4744]: E1201 08:19:30.284599 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:30 crc kubenswrapper[4744]: E1201 08:19:30.284794 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:30 crc kubenswrapper[4744]: E1201 08:19:30.284892 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:30 crc kubenswrapper[4744]: E1201 08:19:30.284956 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.285695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.285746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.285765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.285786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.285804 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.389010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.389072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.389091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.389118 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.389175 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.492113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.492174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.492191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.492219 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.492237 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.595213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.595274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.595292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.595320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.595340 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.698584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.698632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.698649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.698673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.698691 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.802014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.802063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.802081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.802104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.802120 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.904770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.904828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.904847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.904878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:30 crc kubenswrapper[4744]: I1201 08:19:30.904901 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:30Z","lastTransitionTime":"2025-12-01T08:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.007818 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.007875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.007892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.007917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.007936 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.110456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.110509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.110527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.110550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.110568 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.216694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.216752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.216768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.216791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.216810 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.227656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.227708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.227725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.227745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.227761 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: E1201 08:19:31.244741 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.249082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.249123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.249135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.249154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.249166 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: E1201 08:19:31.261800 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.267009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.267051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.267062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.267083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.267095 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: E1201 08:19:31.287339 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.291768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.291812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.291827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.291842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.291855 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: E1201 08:19:31.309401 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.314867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.314924 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.314938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.314957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.314969 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: E1201 08:19:31.331302 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:19:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eae1d0fe-cf0c-4849-8221-3d8cb97835e0\\\",\\\"systemUUID\\\":\\\"734ce460-7574-4e1f-8e70-e2470c8ff4bb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:19:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:19:31 crc kubenswrapper[4744]: E1201 08:19:31.331480 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.333373 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.333547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.333572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.333629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.333648 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.436337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.436398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.436448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.436474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.436492 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.546153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.546222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.546244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.546275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.546297 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.649694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.649811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.649842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.649885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.649913 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.753817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.753888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.753906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.753933 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.753951 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.856797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.856875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.856897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.856928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.856946 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.959457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.959528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.959548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.959574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:31 crc kubenswrapper[4744]: I1201 08:19:31.959594 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:31Z","lastTransitionTime":"2025-12-01T08:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.063505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.063597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.063618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.063650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.063668 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.167888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.167986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.168036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.168072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.168125 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.224524 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:32 crc kubenswrapper[4744]: E1201 08:19:32.224810 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:19:32 crc kubenswrapper[4744]: E1201 08:19:32.224941 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs podName:c8ac81c1-4bf7-472c-8110-21364dde7fd6 nodeName:}" failed. No retries permitted until 2025-12-01 08:20:36.224906072 +0000 UTC m=+168.213964033 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs") pod "network-metrics-daemon-p489k" (UID: "c8ac81c1-4bf7-472c-8110-21364dde7fd6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.272023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.272109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.272138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.272178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.272207 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.285829 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.285939 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.285947 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.285867 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:32 crc kubenswrapper[4744]: E1201 08:19:32.286128 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:32 crc kubenswrapper[4744]: E1201 08:19:32.286268 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:32 crc kubenswrapper[4744]: E1201 08:19:32.286671 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:32 crc kubenswrapper[4744]: E1201 08:19:32.286704 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.376859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.376929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.376948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.376983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.377008 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.480546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.480612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.480635 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.480670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.480693 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.582904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.582963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.582981 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.583006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.583027 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.686247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.686374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.686401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.686510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.686571 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.790244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.790306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.790323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.790348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.790366 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.893664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.893726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.893743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.893767 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.893785 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.997073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.997131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.997140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.997160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:32 crc kubenswrapper[4744]: I1201 08:19:32.997211 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:32Z","lastTransitionTime":"2025-12-01T08:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.099819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.099887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.099906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.099933 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.099956 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.203381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.203458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.203472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.203493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.203508 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.307278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.307361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.307384 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.307440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.307518 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.410186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.410248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.410265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.410290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.410308 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.513167 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.513231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.513248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.513277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.513296 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.617404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.617481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.617492 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.617543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.617558 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.720597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.720673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.720697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.720733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.720756 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.824344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.824475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.824503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.824539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.824565 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.927950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.928043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.928052 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.928076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:33 crc kubenswrapper[4744]: I1201 08:19:33.928088 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:33Z","lastTransitionTime":"2025-12-01T08:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.031293 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.031371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.031389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.031446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.031470 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.134750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.134930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.134946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.135000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.135020 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.238715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.238798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.238817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.238845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.238865 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.284563 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.284657 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.284689 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:34 crc kubenswrapper[4744]: E1201 08:19:34.284749 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:34 crc kubenswrapper[4744]: E1201 08:19:34.284903 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.284936 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:34 crc kubenswrapper[4744]: E1201 08:19:34.292786 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:34 crc kubenswrapper[4744]: E1201 08:19:34.294621 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.341623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.341692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.341710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.341735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.341753 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.445535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.445611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.445632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.445661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.445687 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.548678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.548733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.548750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.548774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.548791 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.652088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.652166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.652184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.652213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.652230 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.755899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.755987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.756012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.756046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.756072 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.859377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.859480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.859498 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.859523 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.859541 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.961877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.961953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.961970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.961993 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:34 crc kubenswrapper[4744]: I1201 08:19:34.962010 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:34Z","lastTransitionTime":"2025-12-01T08:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.064494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.064550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.064565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.064593 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.064613 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.167640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.167694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.167715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.167742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.167762 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.270561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.270636 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.270659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.270690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.270719 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.374475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.374553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.374576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.374602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.374620 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.476680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.476770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.476792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.476819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.476838 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.579502 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.579588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.579610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.579641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.579667 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.681889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.681966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.681988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.682020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.682044 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.784561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.784634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.784654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.784685 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.784703 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.887831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.887892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.887909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.887934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.887954 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.990946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.991017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.991039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.991067 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:35 crc kubenswrapper[4744]: I1201 08:19:35.991090 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:35Z","lastTransitionTime":"2025-12-01T08:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.094779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.094894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.094916 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.094941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.094958 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.197831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.197925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.197945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.198004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.198025 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.284532 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.284573 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.284683 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.284746 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:36 crc kubenswrapper[4744]: E1201 08:19:36.284825 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:36 crc kubenswrapper[4744]: E1201 08:19:36.284960 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:36 crc kubenswrapper[4744]: E1201 08:19:36.285108 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:36 crc kubenswrapper[4744]: E1201 08:19:36.285309 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.301252 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.301306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.301324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.301345 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.301363 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.404939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.405069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.405091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.405115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.405131 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.507872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.507945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.507971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.508001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.508021 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.610974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.611054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.611080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.611109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.611128 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.720004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.720082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.720102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.720127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.720143 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.822766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.822853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.822877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.822904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.822926 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.925590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.925638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.925650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.925668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:36 crc kubenswrapper[4744]: I1201 08:19:36.925680 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:36Z","lastTransitionTime":"2025-12-01T08:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.028945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.028997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.029013 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.029038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.029054 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.132235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.132323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.132342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.132367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.132386 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.234932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.235014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.235034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.235063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.235085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.337779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.337842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.337860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.337887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.337906 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.440735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.440798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.440815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.440841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.440858 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.544009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.544203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.544228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.544262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.544285 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.646913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.646982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.646998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.647023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.647041 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.750574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.750679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.750706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.750736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.750760 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.853829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.853903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.853919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.853942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.853959 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.958278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.958343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.958361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.958387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:37 crc kubenswrapper[4744]: I1201 08:19:37.958446 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:37Z","lastTransitionTime":"2025-12-01T08:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.061329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.061403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.061451 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.061479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.061500 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.164382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.164473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.164490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.164516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.164533 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.268222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.268287 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.268306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.268334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.268352 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.284951 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.284986 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:38 crc kubenswrapper[4744]: E1201 08:19:38.285361 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:38 crc kubenswrapper[4744]: E1201 08:19:38.285544 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.285566 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.285631 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:38 crc kubenswrapper[4744]: E1201 08:19:38.285893 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:38 crc kubenswrapper[4744]: E1201 08:19:38.285956 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.343948 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ncr54" podStartSLOduration=85.343919632 podStartE2EDuration="1m25.343919632s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.343896092 +0000 UTC m=+110.332954053" watchObservedRunningTime="2025-12-01 08:19:38.343919632 +0000 UTC m=+110.332977593" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.368890 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vsh4r" podStartSLOduration=85.368864049 podStartE2EDuration="1m25.368864049s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.368808667 +0000 UTC m=+110.357866618" watchObservedRunningTime="2025-12-01 08:19:38.368864049 +0000 UTC m=+110.357922020" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.374738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.374797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.374815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.374838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.374861 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.457207 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-z4dh2" podStartSLOduration=85.457184835 podStartE2EDuration="1m25.457184835s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.432519176 +0000 UTC m=+110.421577117" watchObservedRunningTime="2025-12-01 08:19:38.457184835 +0000 UTC m=+110.446242796" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.478232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.478282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.478298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.478321 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.478366 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.521494 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podStartSLOduration=85.521476199 podStartE2EDuration="1m25.521476199s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.518739983 +0000 UTC m=+110.507797914" watchObservedRunningTime="2025-12-01 08:19:38.521476199 +0000 UTC m=+110.510534120" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.549984 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lwkvc" podStartSLOduration=85.549959124 podStartE2EDuration="1m25.549959124s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.53511555 +0000 UTC m=+110.524173481" watchObservedRunningTime="2025-12-01 08:19:38.549959124 +0000 UTC m=+110.539017055" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.550495 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=41.550488329 podStartE2EDuration="41.550488329s" podCreationTimestamp="2025-12-01 08:18:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.550261633 +0000 UTC m=+110.539319584" watchObservedRunningTime="2025-12-01 08:19:38.550488329 +0000 UTC m=+110.539546270" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.571428 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=92.571396173 podStartE2EDuration="1m32.571396173s" podCreationTimestamp="2025-12-01 08:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.570972351 +0000 UTC m=+110.560030282" watchObservedRunningTime="2025-12-01 08:19:38.571396173 +0000 UTC m=+110.560454094" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.580977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.581072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.581096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.581128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.581151 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.611552 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=58.611528923 podStartE2EDuration="58.611528923s" podCreationTimestamp="2025-12-01 08:18:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.58811888 +0000 UTC m=+110.577176831" watchObservedRunningTime="2025-12-01 08:19:38.611528923 +0000 UTC m=+110.600586854" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.644772 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=20.644753861 podStartE2EDuration="20.644753861s" podCreationTimestamp="2025-12-01 08:19:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.644594167 +0000 UTC m=+110.633652098" watchObservedRunningTime="2025-12-01 08:19:38.644753861 +0000 UTC m=+110.633811792" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.683158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.683204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.683216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.683235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.683247 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.689837 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.689818169 podStartE2EDuration="1m32.689818169s" podCreationTimestamp="2025-12-01 08:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.677135065 +0000 UTC m=+110.666193006" watchObservedRunningTime="2025-12-01 08:19:38.689818169 +0000 UTC m=+110.678876100" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.712087 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2cvbp" podStartSLOduration=85.712069101 podStartE2EDuration="1m25.712069101s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:38.708562573 +0000 UTC m=+110.697620524" watchObservedRunningTime="2025-12-01 08:19:38.712069101 +0000 UTC m=+110.701127032" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.785963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.786033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.786050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.786076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.786092 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.888731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.888787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.888806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.888829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.888846 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.991164 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.991237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.991256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.991279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:38 crc kubenswrapper[4744]: I1201 08:19:38.991297 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:38Z","lastTransitionTime":"2025-12-01T08:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.094883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.094947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.094971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.095006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.095029 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.197883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.198299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.198482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.198773 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.198935 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.302266 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.302333 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.302350 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.302374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.302392 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.405130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.405176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.405188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.405204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.405213 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.508534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.508621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.508644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.508678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.508700 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.611519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.611594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.611619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.611681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.611706 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.714386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.714475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.714493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.714520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.714536 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.817817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.817884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.817904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.817937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.817960 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.920944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.921010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.921026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.921111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:39 crc kubenswrapper[4744]: I1201 08:19:39.921135 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:39Z","lastTransitionTime":"2025-12-01T08:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.024443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.024506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.024522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.024549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.024566 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.128144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.128242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.128264 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.128290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.128308 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.231592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.231691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.231712 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.231738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.231755 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.285019 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.285088 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.285119 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:40 crc kubenswrapper[4744]: E1201 08:19:40.285706 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:40 crc kubenswrapper[4744]: E1201 08:19:40.285954 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:40 crc kubenswrapper[4744]: E1201 08:19:40.286045 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.285561 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.286195 4744 scope.go:117] "RemoveContainer" containerID="7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1" Dec 01 08:19:40 crc kubenswrapper[4744]: E1201 08:19:40.286823 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:40 crc kubenswrapper[4744]: E1201 08:19:40.286925 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2sflk_openshift-ovn-kubernetes(1df124eb-0511-4702-85d8-3e324c59bb0d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.334504 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.334589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.334606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.334630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.334650 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.438172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.438634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.438791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.438954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.439111 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.542704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.542767 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.542784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.542810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.542827 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.645614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.645679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.645698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.645723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.645742 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.748668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.748987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.749201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.749348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.749524 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.853051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.853329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.853515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.853809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.853963 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.957547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.957633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.957652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.957687 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:40 crc kubenswrapper[4744]: I1201 08:19:40.957726 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:40Z","lastTransitionTime":"2025-12-01T08:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.059909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.059949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.059958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.059974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.059984 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.163555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.164508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.164663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.164807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.164944 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.273231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.273569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.273676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.273772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.273854 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.377437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.377522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.377540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.377568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.377585 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.480547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.481043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.481324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.481709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.482018 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.585948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.586320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.586506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.586669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.586801 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.689664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.690110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.690306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.690526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.690696 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.713464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.713753 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.713940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.714102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.714243 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:19:41Z","lastTransitionTime":"2025-12-01T08:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.788919 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq"] Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.789895 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.793401 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.795980 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.796196 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.796267 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.878865 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/149f41e8-2117-402b-9fcd-43dc1df47adc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.879046 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/149f41e8-2117-402b-9fcd-43dc1df47adc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.879114 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/149f41e8-2117-402b-9fcd-43dc1df47adc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.879149 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/149f41e8-2117-402b-9fcd-43dc1df47adc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.879298 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/149f41e8-2117-402b-9fcd-43dc1df47adc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.980248 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/149f41e8-2117-402b-9fcd-43dc1df47adc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.980386 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/149f41e8-2117-402b-9fcd-43dc1df47adc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.980464 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/149f41e8-2117-402b-9fcd-43dc1df47adc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.980497 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/149f41e8-2117-402b-9fcd-43dc1df47adc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.980667 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/149f41e8-2117-402b-9fcd-43dc1df47adc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.980581 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/149f41e8-2117-402b-9fcd-43dc1df47adc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.980868 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/149f41e8-2117-402b-9fcd-43dc1df47adc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.984988 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/149f41e8-2117-402b-9fcd-43dc1df47adc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:41 crc kubenswrapper[4744]: I1201 08:19:41.990571 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/149f41e8-2117-402b-9fcd-43dc1df47adc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.008329 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/149f41e8-2117-402b-9fcd-43dc1df47adc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sp2dq\" (UID: \"149f41e8-2117-402b-9fcd-43dc1df47adc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.114034 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.284048 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.284166 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:42 crc kubenswrapper[4744]: E1201 08:19:42.284224 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.284080 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:42 crc kubenswrapper[4744]: E1201 08:19:42.284337 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.284278 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:42 crc kubenswrapper[4744]: E1201 08:19:42.284502 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:42 crc kubenswrapper[4744]: E1201 08:19:42.284585 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.980061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" event={"ID":"149f41e8-2117-402b-9fcd-43dc1df47adc","Type":"ContainerStarted","Data":"81ce98bb5322762ee763fb325780a71eb4faad84057b44288d62245318561b21"} Dec 01 08:19:42 crc kubenswrapper[4744]: I1201 08:19:42.980144 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" event={"ID":"149f41e8-2117-402b-9fcd-43dc1df47adc","Type":"ContainerStarted","Data":"b2ce15eff7c7e3be35c4cf0a31f0c75dd8da34008bfc2e6dcf77c73738ddde4c"} Dec 01 08:19:44 crc kubenswrapper[4744]: I1201 08:19:44.284653 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:44 crc kubenswrapper[4744]: I1201 08:19:44.284708 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:44 crc kubenswrapper[4744]: I1201 08:19:44.284775 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:44 crc kubenswrapper[4744]: E1201 08:19:44.284821 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:44 crc kubenswrapper[4744]: I1201 08:19:44.284847 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:44 crc kubenswrapper[4744]: E1201 08:19:44.284978 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:44 crc kubenswrapper[4744]: E1201 08:19:44.285125 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:44 crc kubenswrapper[4744]: E1201 08:19:44.285336 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:46 crc kubenswrapper[4744]: I1201 08:19:46.284715 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:46 crc kubenswrapper[4744]: I1201 08:19:46.284762 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:46 crc kubenswrapper[4744]: I1201 08:19:46.284878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:46 crc kubenswrapper[4744]: I1201 08:19:46.284873 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:46 crc kubenswrapper[4744]: E1201 08:19:46.284877 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:46 crc kubenswrapper[4744]: E1201 08:19:46.285046 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:46 crc kubenswrapper[4744]: E1201 08:19:46.285142 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:46 crc kubenswrapper[4744]: E1201 08:19:46.285203 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.004137 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/1.log" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.004932 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/0.log" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.005013 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a5e5532-8675-4f3f-aa30-7c22bc95b311" containerID="e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639" exitCode=1 Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.005108 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerDied","Data":"e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639"} Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.005177 4744 scope.go:117] "RemoveContainer" containerID="49002ba120d3284b4758fdcb8a481709ea5cd653ca8f089a5df07c5118950bb9" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.005820 4744 scope.go:117] "RemoveContainer" containerID="e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639" Dec 01 08:19:48 crc kubenswrapper[4744]: E1201 08:19:48.006176 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vsh4r_openshift-multus(6a5e5532-8675-4f3f-aa30-7c22bc95b311)\"" pod="openshift-multus/multus-vsh4r" podUID="6a5e5532-8675-4f3f-aa30-7c22bc95b311" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.033294 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sp2dq" podStartSLOduration=95.033269328 podStartE2EDuration="1m35.033269328s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:43.001176337 +0000 UTC m=+114.990234298" watchObservedRunningTime="2025-12-01 08:19:48.033269328 +0000 UTC m=+120.022327289" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.284047 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.284124 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.284206 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:48 crc kubenswrapper[4744]: E1201 08:19:48.284385 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:48 crc kubenswrapper[4744]: I1201 08:19:48.284449 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:48 crc kubenswrapper[4744]: E1201 08:19:48.286723 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:48 crc kubenswrapper[4744]: E1201 08:19:48.286888 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:48 crc kubenswrapper[4744]: E1201 08:19:48.286951 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:48 crc kubenswrapper[4744]: E1201 08:19:48.301778 4744 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 01 08:19:48 crc kubenswrapper[4744]: E1201 08:19:48.403498 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 08:19:49 crc kubenswrapper[4744]: I1201 08:19:49.010478 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/1.log" Dec 01 08:19:50 crc kubenswrapper[4744]: I1201 08:19:50.284534 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:50 crc kubenswrapper[4744]: I1201 08:19:50.284672 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:50 crc kubenswrapper[4744]: I1201 08:19:50.284758 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:50 crc kubenswrapper[4744]: E1201 08:19:50.284917 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:50 crc kubenswrapper[4744]: I1201 08:19:50.285221 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:50 crc kubenswrapper[4744]: E1201 08:19:50.285329 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:50 crc kubenswrapper[4744]: E1201 08:19:50.285678 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:50 crc kubenswrapper[4744]: E1201 08:19:50.285872 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:52 crc kubenswrapper[4744]: I1201 08:19:52.284308 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:52 crc kubenswrapper[4744]: I1201 08:19:52.284399 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:52 crc kubenswrapper[4744]: I1201 08:19:52.284344 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:52 crc kubenswrapper[4744]: I1201 08:19:52.284302 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:52 crc kubenswrapper[4744]: E1201 08:19:52.284537 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:52 crc kubenswrapper[4744]: E1201 08:19:52.284664 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:52 crc kubenswrapper[4744]: E1201 08:19:52.285251 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:52 crc kubenswrapper[4744]: E1201 08:19:52.285490 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:52 crc kubenswrapper[4744]: I1201 08:19:52.285713 4744 scope.go:117] "RemoveContainer" containerID="7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1" Dec 01 08:19:53 crc kubenswrapper[4744]: I1201 08:19:53.027879 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/3.log" Dec 01 08:19:53 crc kubenswrapper[4744]: I1201 08:19:53.031993 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerStarted","Data":"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b"} Dec 01 08:19:53 crc kubenswrapper[4744]: I1201 08:19:53.032596 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:19:53 crc kubenswrapper[4744]: I1201 08:19:53.084359 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podStartSLOduration=100.084335897 podStartE2EDuration="1m40.084335897s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:19:53.083464823 +0000 UTC m=+125.072522814" watchObservedRunningTime="2025-12-01 08:19:53.084335897 +0000 UTC m=+125.073393848" Dec 01 08:19:53 crc kubenswrapper[4744]: I1201 08:19:53.273178 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p489k"] Dec 01 08:19:53 crc kubenswrapper[4744]: I1201 08:19:53.273349 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:53 crc kubenswrapper[4744]: E1201 08:19:53.273519 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:53 crc kubenswrapper[4744]: E1201 08:19:53.404534 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 08:19:54 crc kubenswrapper[4744]: I1201 08:19:54.283983 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:54 crc kubenswrapper[4744]: I1201 08:19:54.284015 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:54 crc kubenswrapper[4744]: I1201 08:19:54.284086 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:54 crc kubenswrapper[4744]: E1201 08:19:54.284672 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:54 crc kubenswrapper[4744]: E1201 08:19:54.284544 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:54 crc kubenswrapper[4744]: E1201 08:19:54.284822 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:55 crc kubenswrapper[4744]: I1201 08:19:55.284831 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:55 crc kubenswrapper[4744]: E1201 08:19:55.285050 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:56 crc kubenswrapper[4744]: I1201 08:19:56.284243 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:56 crc kubenswrapper[4744]: I1201 08:19:56.284282 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:56 crc kubenswrapper[4744]: E1201 08:19:56.284516 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:56 crc kubenswrapper[4744]: I1201 08:19:56.284614 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:56 crc kubenswrapper[4744]: E1201 08:19:56.284841 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:56 crc kubenswrapper[4744]: E1201 08:19:56.284954 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:57 crc kubenswrapper[4744]: I1201 08:19:57.284506 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:57 crc kubenswrapper[4744]: E1201 08:19:57.284694 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:19:58 crc kubenswrapper[4744]: I1201 08:19:58.284239 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:19:58 crc kubenswrapper[4744]: I1201 08:19:58.284232 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:19:58 crc kubenswrapper[4744]: I1201 08:19:58.284319 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:19:58 crc kubenswrapper[4744]: E1201 08:19:58.286848 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:19:58 crc kubenswrapper[4744]: E1201 08:19:58.287019 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:19:58 crc kubenswrapper[4744]: E1201 08:19:58.287247 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:19:58 crc kubenswrapper[4744]: E1201 08:19:58.406352 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 08:19:59 crc kubenswrapper[4744]: I1201 08:19:59.284974 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:19:59 crc kubenswrapper[4744]: E1201 08:19:59.285185 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:20:00 crc kubenswrapper[4744]: I1201 08:20:00.284927 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:00 crc kubenswrapper[4744]: I1201 08:20:00.284942 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:00 crc kubenswrapper[4744]: E1201 08:20:00.285145 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:20:00 crc kubenswrapper[4744]: E1201 08:20:00.285246 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:20:00 crc kubenswrapper[4744]: I1201 08:20:00.285252 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:20:00 crc kubenswrapper[4744]: E1201 08:20:00.285349 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:20:01 crc kubenswrapper[4744]: I1201 08:20:01.284857 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:20:01 crc kubenswrapper[4744]: E1201 08:20:01.285358 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:20:01 crc kubenswrapper[4744]: I1201 08:20:01.285771 4744 scope.go:117] "RemoveContainer" containerID="e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639" Dec 01 08:20:02 crc kubenswrapper[4744]: I1201 08:20:02.068588 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/1.log" Dec 01 08:20:02 crc kubenswrapper[4744]: I1201 08:20:02.068956 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerStarted","Data":"c5cd4b8744689d1838bd2d0e5d0166485469706e4e1b6d8f26489c876c06341e"} Dec 01 08:20:02 crc kubenswrapper[4744]: I1201 08:20:02.285218 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:02 crc kubenswrapper[4744]: I1201 08:20:02.285324 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:02 crc kubenswrapper[4744]: I1201 08:20:02.285361 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:20:02 crc kubenswrapper[4744]: E1201 08:20:02.285546 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:20:02 crc kubenswrapper[4744]: E1201 08:20:02.285697 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:20:02 crc kubenswrapper[4744]: E1201 08:20:02.285791 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:20:03 crc kubenswrapper[4744]: I1201 08:20:03.284262 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:20:03 crc kubenswrapper[4744]: E1201 08:20:03.284546 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p489k" podUID="c8ac81c1-4bf7-472c-8110-21364dde7fd6" Dec 01 08:20:04 crc kubenswrapper[4744]: I1201 08:20:04.284453 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:20:04 crc kubenswrapper[4744]: I1201 08:20:04.284507 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:04 crc kubenswrapper[4744]: I1201 08:20:04.284467 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:04 crc kubenswrapper[4744]: I1201 08:20:04.287333 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 08:20:04 crc kubenswrapper[4744]: I1201 08:20:04.288946 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 08:20:04 crc kubenswrapper[4744]: I1201 08:20:04.289187 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 08:20:04 crc kubenswrapper[4744]: I1201 08:20:04.289259 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 08:20:05 crc kubenswrapper[4744]: I1201 08:20:05.284189 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:20:05 crc kubenswrapper[4744]: I1201 08:20:05.287724 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 08:20:05 crc kubenswrapper[4744]: I1201 08:20:05.288646 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.392388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.497729 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dk5mh"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.498484 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.505534 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.506629 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.510163 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wl2vv"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.511318 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.511734 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.512181 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.517849 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.517948 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6qsl"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.520005 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.520296 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.520965 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.521285 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.521498 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.521744 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.521858 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.522019 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.522093 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.522137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.522965 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.524227 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mvtcw"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.524877 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.527608 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-l9mth"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.528288 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.530406 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.530667 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.530874 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.530908 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531130 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531221 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531400 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531588 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531651 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531748 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531825 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531901 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531925 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.531595 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.532086 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.532153 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.532177 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.532799 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.533557 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.534692 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.546942 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.548049 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.548256 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.548612 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.548690 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.548945 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pflgv"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.549025 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.549177 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.550114 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.558554 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.558808 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.561707 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.561885 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.562240 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.562254 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.574016 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.574158 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.574380 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.574665 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.575805 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a368aebd-97fc-44cf-bd2b-755850a3b5d2-serving-cert\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.575848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2w22\" (UniqueName: \"kubernetes.io/projected/a368aebd-97fc-44cf-bd2b-755850a3b5d2-kube-api-access-z2w22\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.575878 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a368aebd-97fc-44cf-bd2b-755850a3b5d2-config\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.575990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a368aebd-97fc-44cf-bd2b-755850a3b5d2-trusted-ca\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576115 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mq727"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576136 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576197 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576315 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576409 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576433 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576499 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576519 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576323 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576588 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576608 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mq727" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576704 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576742 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576806 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576931 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576948 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576964 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.576705 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.577108 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.577117 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.577195 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.577424 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.579909 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l6djk"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.580191 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.580355 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.581439 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.582143 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.582511 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.582958 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.583076 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-nmgd9"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.583648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.585052 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pnhzx"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.586747 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.587988 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.588122 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.588247 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.588317 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.588632 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.589048 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.589588 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.589068 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.590963 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.592499 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.592673 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.592927 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.593286 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.593374 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.593662 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.593963 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.596302 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-955gc"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.596413 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.596762 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.597697 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.597846 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.598077 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.601548 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.602046 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.605109 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44bgh"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.605653 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.605734 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.606314 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.606705 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.606805 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.607240 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.609739 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-q6bv4"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.615112 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.615443 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.615635 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.616199 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.616393 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.616633 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.619076 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.620254 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.620680 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.621073 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.621090 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.621870 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.622241 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.622410 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.622736 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.623202 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.634754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.643840 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.644281 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.644700 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.646011 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rhrgq"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.646230 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.646929 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.651095 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.651105 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.651560 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.652619 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qpjtj"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.652809 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.653561 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.654355 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.654528 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.659310 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.659728 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.660339 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.660515 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.661116 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.661550 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.662193 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.662380 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.662447 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.662452 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.662511 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.663106 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nl48d"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.663504 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.663770 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.663906 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.664237 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.664524 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.664562 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5kq46"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.664778 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665072 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dk5mh"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665089 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-l9mth"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665099 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6qsl"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665107 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mq727"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665159 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665319 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665412 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.665482 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.666191 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.667665 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l6djk"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.668076 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mvtcw"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.669043 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.670031 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wl2vv"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.670882 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rhrgq"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.671796 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.673353 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.674499 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676275 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pflgv"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676683 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-audit-dir\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676707 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-config\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41c7aa1f-b684-4370-ab59-87aeb78251cd-config\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676751 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a368aebd-97fc-44cf-bd2b-755850a3b5d2-trusted-ca\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676778 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676796 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-audit\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676825 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv997\" (UniqueName: \"kubernetes.io/projected/5b7ea6bf-6f2c-483e-b37b-9032d56018ee-kube-api-access-cv997\") pod \"downloads-7954f5f757-mq727\" (UID: \"5b7ea6bf-6f2c-483e-b37b-9032d56018ee\") " pod="openshift-console/downloads-7954f5f757-mq727" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676843 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4bb8\" (UniqueName: \"kubernetes.io/projected/268ee28c-0643-4161-aee8-6fa5a08722c8-kube-api-access-q4bb8\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.676856 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/214a177a-1cc2-4322-8994-400d072af8e7-etcd-client\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.678595 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.678667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.678953 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-client-ca\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679233 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr2p4\" (UniqueName: \"kubernetes.io/projected/214a177a-1cc2-4322-8994-400d072af8e7-kube-api-access-zr2p4\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679381 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b80c9c0b-07a4-441e-9f71-f9962901967e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679419 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679443 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679470 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679498 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa-metrics-tls\") pod \"dns-operator-744455d44c-pnhzx\" (UID: \"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.679946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbjd9\" (UniqueName: \"kubernetes.io/projected/41c7aa1f-b684-4370-ab59-87aeb78251cd-kube-api-access-jbjd9\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2w22\" (UniqueName: \"kubernetes.io/projected/a368aebd-97fc-44cf-bd2b-755850a3b5d2-kube-api-access-z2w22\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680309 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29rjx\" (UniqueName: \"kubernetes.io/projected/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-kube-api-access-29rjx\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680329 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54063efd-02f0-46a6-8f06-edd7070fda31-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680349 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87bf7b3-7b9d-407d-a113-387afac58cab-serving-cert\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680376 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80c9c0b-07a4-441e-9f71-f9962901967e-config\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680395 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f71e09c5-239a-4982-ab7d-93c6d647048a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9rr8l\" (UID: \"f71e09c5-239a-4982-ab7d-93c6d647048a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680447 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-service-ca\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/268ee28c-0643-4161-aee8-6fa5a08722c8-serving-cert\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-encryption-config\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680528 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-trusted-ca-bundle\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680550 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n65ff\" (UniqueName: \"kubernetes.io/projected/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-kube-api-access-n65ff\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7kmg\" (UniqueName: \"kubernetes.io/projected/f71e09c5-239a-4982-ab7d-93c6d647048a-kube-api-access-s7kmg\") pod \"cluster-samples-operator-665b6dd947-9rr8l\" (UID: \"f71e09c5-239a-4982-ab7d-93c6d647048a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-image-import-ca\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680757 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8cn7\" (UniqueName: \"kubernetes.io/projected/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-kube-api-access-c8cn7\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-config\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.680956 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681221 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-etcd-ca\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681327 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681349 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce21f093-ef49-49f7-b8aa-e1159c94b30e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681365 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-serving-cert\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-policies\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681410 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/214a177a-1cc2-4322-8994-400d072af8e7-serving-cert\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-config\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681536 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd87x\" (UniqueName: \"kubernetes.io/projected/a87bf7b3-7b9d-407d-a113-387afac58cab-kube-api-access-rd87x\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681630 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54063efd-02f0-46a6-8f06-edd7070fda31-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-etcd-service-ca\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-serving-cert\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681716 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-config\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681748 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7qjq\" (UniqueName: \"kubernetes.io/projected/ce21f093-ef49-49f7-b8aa-e1159c94b30e-kube-api-access-b7qjq\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681776 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-client-ca\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-oauth-serving-cert\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-dir\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681867 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a368aebd-97fc-44cf-bd2b-755850a3b5d2-serving-cert\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681894 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrl6v\" (UniqueName: \"kubernetes.io/projected/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-kube-api-access-mrl6v\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681925 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a368aebd-97fc-44cf-bd2b-755850a3b5d2-config\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681948 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-oauth-config\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.681981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-images\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682012 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-config\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682041 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682074 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-serving-cert\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682097 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682126 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/41c7aa1f-b684-4370-ab59-87aeb78251cd-auth-proxy-config\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682156 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682189 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce21f093-ef49-49f7-b8aa-e1159c94b30e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-config\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682242 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b80c9c0b-07a4-441e-9f71-f9962901967e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682294 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-node-pullsecrets\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682317 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-etcd-client\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682366 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-config\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682389 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-service-ca-bundle\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682469 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vclbj\" (UniqueName: \"kubernetes.io/projected/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-kube-api-access-vclbj\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682496 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-config\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52tlf\" (UniqueName: \"kubernetes.io/projected/b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa-kube-api-access-52tlf\") pod \"dns-operator-744455d44c-pnhzx\" (UID: \"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682545 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a368aebd-97fc-44cf-bd2b-755850a3b5d2-trusted-ca\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-etcd-serving-ca\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682597 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54063efd-02f0-46a6-8f06-edd7070fda31-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682624 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/41c7aa1f-b684-4370-ab59-87aeb78251cd-machine-approver-tls\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.682749 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a368aebd-97fc-44cf-bd2b-755850a3b5d2-config\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.683639 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.686809 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.689603 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.690776 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nmgd9"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.691830 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pnhzx"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.692942 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-955gc"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.694328 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.695806 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.696083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a368aebd-97fc-44cf-bd2b-755850a3b5d2-serving-cert\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.697325 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qpjtj"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.698387 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.700080 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.700742 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.701024 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.702075 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.703123 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nl48d"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.704208 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.705462 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.706665 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.707598 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wr5kd"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.709028 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.710321 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.713502 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44bgh"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.716982 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.722086 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.726911 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-wvscw"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.727806 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.731993 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-lj9cj"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.732780 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.734194 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wr5kd"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.735455 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-wvscw"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.736792 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5kq46"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.737945 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-ml6wn"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.738811 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.739295 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ml6wn"] Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.741168 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.761257 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa-metrics-tls\") pod \"dns-operator-744455d44c-pnhzx\" (UID: \"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbjd9\" (UniqueName: \"kubernetes.io/projected/41c7aa1f-b684-4370-ab59-87aeb78251cd-kube-api-access-jbjd9\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29rjx\" (UniqueName: \"kubernetes.io/projected/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-kube-api-access-29rjx\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54063efd-02f0-46a6-8f06-edd7070fda31-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87bf7b3-7b9d-407d-a113-387afac58cab-serving-cert\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80c9c0b-07a4-441e-9f71-f9962901967e-config\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f71e09c5-239a-4982-ab7d-93c6d647048a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9rr8l\" (UID: \"f71e09c5-239a-4982-ab7d-93c6d647048a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783384 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-service-ca\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783411 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/268ee28c-0643-4161-aee8-6fa5a08722c8-serving-cert\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-encryption-config\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-trusted-ca-bundle\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783501 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n65ff\" (UniqueName: \"kubernetes.io/projected/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-kube-api-access-n65ff\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783526 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7kmg\" (UniqueName: \"kubernetes.io/projected/f71e09c5-239a-4982-ab7d-93c6d647048a-kube-api-access-s7kmg\") pod \"cluster-samples-operator-665b6dd947-9rr8l\" (UID: \"f71e09c5-239a-4982-ab7d-93c6d647048a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783550 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783574 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-image-import-ca\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783601 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8cn7\" (UniqueName: \"kubernetes.io/projected/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-kube-api-access-c8cn7\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-config\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783686 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-etcd-ca\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783709 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783731 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783755 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce21f093-ef49-49f7-b8aa-e1159c94b30e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783775 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-serving-cert\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783798 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-policies\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/214a177a-1cc2-4322-8994-400d072af8e7-serving-cert\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783848 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-config\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd87x\" (UniqueName: \"kubernetes.io/projected/a87bf7b3-7b9d-407d-a113-387afac58cab-kube-api-access-rd87x\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54063efd-02f0-46a6-8f06-edd7070fda31-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-etcd-service-ca\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.783980 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-serving-cert\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-config\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7qjq\" (UniqueName: \"kubernetes.io/projected/ce21f093-ef49-49f7-b8aa-e1159c94b30e-kube-api-access-b7qjq\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784055 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-client-ca\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-oauth-serving-cert\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-dir\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrl6v\" (UniqueName: \"kubernetes.io/projected/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-kube-api-access-mrl6v\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784156 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-oauth-config\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-images\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-config\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-serving-cert\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/41c7aa1f-b684-4370-ab59-87aeb78251cd-auth-proxy-config\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784372 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce21f093-ef49-49f7-b8aa-e1159c94b30e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-config\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784439 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b80c9c0b-07a4-441e-9f71-f9962901967e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784477 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-node-pullsecrets\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784500 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-etcd-client\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784538 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-config\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784563 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-service-ca-bundle\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784588 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784614 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vclbj\" (UniqueName: \"kubernetes.io/projected/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-kube-api-access-vclbj\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-config\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52tlf\" (UniqueName: \"kubernetes.io/projected/b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa-kube-api-access-52tlf\") pod \"dns-operator-744455d44c-pnhzx\" (UID: \"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784680 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-etcd-serving-ca\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784706 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54063efd-02f0-46a6-8f06-edd7070fda31-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784727 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/41c7aa1f-b684-4370-ab59-87aeb78251cd-machine-approver-tls\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784748 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-audit-dir\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784770 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-config\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.784790 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41c7aa1f-b684-4370-ab59-87aeb78251cd-config\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785361 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785393 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785439 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-audit\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785464 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv997\" (UniqueName: \"kubernetes.io/projected/5b7ea6bf-6f2c-483e-b37b-9032d56018ee-kube-api-access-cv997\") pod \"downloads-7954f5f757-mq727\" (UID: \"5b7ea6bf-6f2c-483e-b37b-9032d56018ee\") " pod="openshift-console/downloads-7954f5f757-mq727" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4bb8\" (UniqueName: \"kubernetes.io/projected/268ee28c-0643-4161-aee8-6fa5a08722c8-kube-api-access-q4bb8\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785485 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-image-import-ca\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/214a177a-1cc2-4322-8994-400d072af8e7-etcd-client\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785731 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785760 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785786 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-client-ca\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785814 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr2p4\" (UniqueName: \"kubernetes.io/projected/214a177a-1cc2-4322-8994-400d072af8e7-kube-api-access-zr2p4\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b80c9c0b-07a4-441e-9f71-f9962901967e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785861 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785883 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.785930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.786227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-trusted-ca-bundle\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.786803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-service-ca\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.786882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87bf7b3-7b9d-407d-a113-387afac58cab-serving-cert\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.787667 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.787837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-config\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.788093 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa-metrics-tls\") pod \"dns-operator-744455d44c-pnhzx\" (UID: \"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.788160 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.788493 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/41c7aa1f-b684-4370-ab59-87aeb78251cd-auth-proxy-config\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.788893 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-encryption-config\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.788907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.789302 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.789959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-config\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.790447 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.790768 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.790855 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.790857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41c7aa1f-b684-4370-ab59-87aeb78251cd-config\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.790956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-etcd-serving-ca\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.791098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f71e09c5-239a-4982-ab7d-93c6d647048a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9rr8l\" (UID: \"f71e09c5-239a-4982-ab7d-93c6d647048a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.791552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-audit\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.791655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.791806 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/268ee28c-0643-4161-aee8-6fa5a08722c8-serving-cert\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.792054 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-serving-cert\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.792116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.792544 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-client-ca\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.792869 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-config\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.792890 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-config\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.792900 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-client-ca\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.793157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-oauth-serving-cert\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.793200 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-dir\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.793201 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.793579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.793757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-audit-dir\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.793904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-policies\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.794006 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.794622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.794860 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-service-ca-bundle\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.794885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-images\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.794955 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-node-pullsecrets\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.795372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-config\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.795460 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-config\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.795631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce21f093-ef49-49f7-b8aa-e1159c94b30e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.796028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/41c7aa1f-b684-4370-ab59-87aeb78251cd-machine-approver-tls\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.796396 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.796724 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.796770 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-serving-cert\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.797406 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-oauth-config\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.797451 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.798213 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-serving-cert\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.798539 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-etcd-client\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.800143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce21f093-ef49-49f7-b8aa-e1159c94b30e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.801181 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.820948 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.827362 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-config\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.841083 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.862002 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.881235 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.890272 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.904103 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.910009 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b80c9c0b-07a4-441e-9f71-f9962901967e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.921862 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.925133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b80c9c0b-07a4-441e-9f71-f9962901967e-config\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.941508 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.963722 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 08:20:12 crc kubenswrapper[4744]: I1201 08:20:12.981119 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.001634 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.009910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/214a177a-1cc2-4322-8994-400d072af8e7-serving-cert\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.021749 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.029920 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-etcd-service-ca\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.041784 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.043674 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.061190 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.069550 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-etcd-ca\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.082276 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.101646 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.121598 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.123041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/214a177a-1cc2-4322-8994-400d072af8e7-config\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.142288 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.151753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/214a177a-1cc2-4322-8994-400d072af8e7-etcd-client\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.161818 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.181963 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.202707 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.222610 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.236798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54063efd-02f0-46a6-8f06-edd7070fda31-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.240943 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.245049 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54063efd-02f0-46a6-8f06-edd7070fda31-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.281147 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.301268 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.321499 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.342452 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.361722 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.382365 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.401441 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.421556 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.442299 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.461201 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.481676 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.501080 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.541823 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.561233 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.581149 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.600816 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.621555 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.641072 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.659459 4744 request.go:700] Waited for 1.012715417s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-certs-default&limit=500&resourceVersion=0 Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.661326 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.681592 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.715204 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.722596 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.741630 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.761783 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.782040 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.802109 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.821312 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.841506 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.862396 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.882003 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.902294 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.921296 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.941858 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.961703 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 08:20:13 crc kubenswrapper[4744]: I1201 08:20:13.981619 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.001746 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.021355 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.042064 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.061976 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.081963 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.101166 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.120762 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.140966 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.161118 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.181771 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.200944 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.205930 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:14 crc kubenswrapper[4744]: E1201 08:20:14.206626 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:22:16.206576406 +0000 UTC m=+268.195634367 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.221388 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.241774 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.261677 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.282891 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.309195 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.310529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.321026 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2w22\" (UniqueName: \"kubernetes.io/projected/a368aebd-97fc-44cf-bd2b-755850a3b5d2-kube-api-access-z2w22\") pod \"console-operator-58897d9998-dk5mh\" (UID: \"a368aebd-97fc-44cf-bd2b-755850a3b5d2\") " pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.321317 4744 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.332291 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.341923 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.362488 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.382025 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.402514 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.410538 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.410631 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.410726 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.416894 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.416944 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.421393 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.422071 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.446518 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.465240 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.482151 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.502291 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.512773 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.521595 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.528963 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.540911 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.541112 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.562210 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.604218 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dk5mh"] Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.608029 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbjd9\" (UniqueName: \"kubernetes.io/projected/41c7aa1f-b684-4370-ab59-87aeb78251cd-kube-api-access-jbjd9\") pod \"machine-approver-56656f9798-hzlbv\" (UID: \"41c7aa1f-b684-4370-ab59-87aeb78251cd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:14 crc kubenswrapper[4744]: W1201 08:20:14.619629 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda368aebd_97fc_44cf_bd2b_755850a3b5d2.slice/crio-86b88b6711a10e738fd925463fc12ff111717b47868900ddce5f08b9e5639531 WatchSource:0}: Error finding container 86b88b6711a10e738fd925463fc12ff111717b47868900ddce5f08b9e5639531: Status 404 returned error can't find the container with id 86b88b6711a10e738fd925463fc12ff111717b47868900ddce5f08b9e5639531 Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.630913 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29rjx\" (UniqueName: \"kubernetes.io/projected/9a9c9a23-151c-41ce-9a1b-57adbd823c1e-kube-api-access-29rjx\") pod \"apiserver-76f77b778f-wl2vv\" (UID: \"9a9c9a23-151c-41ce-9a1b-57adbd823c1e\") " pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.642654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n65ff\" (UniqueName: \"kubernetes.io/projected/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-kube-api-access-n65ff\") pod \"console-f9d7485db-nmgd9\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.660059 4744 request.go:700] Waited for 1.874361264s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication-operator/serviceaccounts/authentication-operator/token Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.663098 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.663580 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7kmg\" (UniqueName: \"kubernetes.io/projected/f71e09c5-239a-4982-ab7d-93c6d647048a-kube-api-access-s7kmg\") pod \"cluster-samples-operator-665b6dd947-9rr8l\" (UID: \"f71e09c5-239a-4982-ab7d-93c6d647048a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.681806 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8cn7\" (UniqueName: \"kubernetes.io/projected/f71710a5-e0fa-4ac7-be6a-80e84a342e3b-kube-api-access-c8cn7\") pod \"authentication-operator-69f744f599-mvtcw\" (UID: \"f71710a5-e0fa-4ac7-be6a-80e84a342e3b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.707230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b80c9c0b-07a4-441e-9f71-f9962901967e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lh52z\" (UID: \"b80c9c0b-07a4-441e-9f71-f9962901967e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.707507 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.717994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr2p4\" (UniqueName: \"kubernetes.io/projected/214a177a-1cc2-4322-8994-400d072af8e7-kube-api-access-zr2p4\") pod \"etcd-operator-b45778765-955gc\" (UID: \"214a177a-1cc2-4322-8994-400d072af8e7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.757088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52tlf\" (UniqueName: \"kubernetes.io/projected/b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa-kube-api-access-52tlf\") pod \"dns-operator-744455d44c-pnhzx\" (UID: \"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.779639 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.786107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrl6v\" (UniqueName: \"kubernetes.io/projected/6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc-kube-api-access-mrl6v\") pod \"machine-api-operator-5694c8668f-l9mth\" (UID: \"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.788687 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.799356 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7qjq\" (UniqueName: \"kubernetes.io/projected/ce21f093-ef49-49f7-b8aa-e1159c94b30e-kube-api-access-b7qjq\") pod \"openshift-apiserver-operator-796bbdcf4f-pw4dg\" (UID: \"ce21f093-ef49-49f7-b8aa-e1159c94b30e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.801141 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.816715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv997\" (UniqueName: \"kubernetes.io/projected/5b7ea6bf-6f2c-483e-b37b-9032d56018ee-kube-api-access-cv997\") pod \"downloads-7954f5f757-mq727\" (UID: \"5b7ea6bf-6f2c-483e-b37b-9032d56018ee\") " pod="openshift-console/downloads-7954f5f757-mq727" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.836641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4bb8\" (UniqueName: \"kubernetes.io/projected/268ee28c-0643-4161-aee8-6fa5a08722c8-kube-api-access-q4bb8\") pod \"route-controller-manager-6576b87f9c-nr58s\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.839358 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.845227 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.853482 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54063efd-02f0-46a6-8f06-edd7070fda31-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whgj6\" (UID: \"54063efd-02f0-46a6-8f06-edd7070fda31\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.873305 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.874483 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd87x\" (UniqueName: \"kubernetes.io/projected/a87bf7b3-7b9d-407d-a113-387afac58cab-kube-api-access-rd87x\") pod \"controller-manager-879f6c89f-r6qsl\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.889188 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" Dec 01 08:20:14 crc kubenswrapper[4744]: I1201 08:20:14.894696 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wl2vv"] Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.124337 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" event={"ID":"41c7aa1f-b684-4370-ab59-87aeb78251cd","Type":"ContainerStarted","Data":"af2fa0bcacf687305e480d6c51f62dd4daf90fa196084655768dd04285601a71"} Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.126131 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dk5mh" event={"ID":"a368aebd-97fc-44cf-bd2b-755850a3b5d2","Type":"ContainerStarted","Data":"86b88b6711a10e738fd925463fc12ff111717b47868900ddce5f08b9e5639531"} Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.176312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.176567 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.176923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vclbj\" (UniqueName: \"kubernetes.io/projected/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-kube-api-access-vclbj\") pod \"oauth-openshift-558db77b4-pflgv\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.177227 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.177271 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.177368 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.178001 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mq727" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.178370 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-bound-sa-token\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.178467 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.178512 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5516996f-81ea-44d0-b5a1-5af61f00af8c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.178568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bwbt\" (UniqueName: \"kubernetes.io/projected/41e0738b-fa8f-425b-8657-452c59c0b8fc-kube-api-access-9bwbt\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.178711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41e0738b-fa8f-425b-8657-452c59c0b8fc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.178901 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.179040 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-trusted-ca\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.179179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnhhk\" (UniqueName: \"kubernetes.io/projected/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-kube-api-access-qnhhk\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.179468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-tls\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.179526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41e0738b-fa8f-425b-8657-452c59c0b8fc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.179562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5516996f-81ea-44d0-b5a1-5af61f00af8c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.179771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.179807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t85fp\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-kube-api-access-t85fp\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.181110 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/41e0738b-fa8f-425b-8657-452c59c0b8fc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.181163 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-certificates\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.181219 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:15.681188627 +0000 UTC m=+147.670246588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.186586 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30a4c368-98eb-4f52-be7b-0d6aa7d50a55-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-k997m\" (UID: \"30a4c368-98eb-4f52-be7b-0d6aa7d50a55\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:15 crc kubenswrapper[4744]: W1201 08:20:15.187220 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-a701a286b2f661f6b89d7466cbb1fa9cd0903a59fc6268723d0efb269efdabf9 WatchSource:0}: Error finding container a701a286b2f661f6b89d7466cbb1fa9cd0903a59fc6268723d0efb269efdabf9: Status 404 returned error can't find the container with id a701a286b2f661f6b89d7466cbb1fa9cd0903a59fc6268723d0efb269efdabf9 Dec 01 08:20:15 crc kubenswrapper[4744]: W1201 08:20:15.194434 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-a0acc4e47f7400679754462b93f3a43cd349bf05099f6e3e11387bec177557dd WatchSource:0}: Error finding container a0acc4e47f7400679754462b93f3a43cd349bf05099f6e3e11387bec177557dd: Status 404 returned error can't find the container with id a0acc4e47f7400679754462b93f3a43cd349bf05099f6e3e11387bec177557dd Dec 01 08:20:15 crc kubenswrapper[4744]: W1201 08:20:15.194801 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-8c79b4a0bf49e55a5ef64084c839ea07b66b4bc23d32e159514ba23634466f14 WatchSource:0}: Error finding container 8c79b4a0bf49e55a5ef64084c839ea07b66b4bc23d32e159514ba23634466f14: Status 404 returned error can't find the container with id 8c79b4a0bf49e55a5ef64084c839ea07b66b4bc23d32e159514ba23634466f14 Dec 01 08:20:15 crc kubenswrapper[4744]: W1201 08:20:15.220851 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a9c9a23_151c_41ce_9a1b_57adbd823c1e.slice/crio-4b0222056e99f1fd8ff216f9fa96a36f3fac6b2f36d9e21b4386e1d58ecad35c WatchSource:0}: Error finding container 4b0222056e99f1fd8ff216f9fa96a36f3fac6b2f36d9e21b4386e1d58ecad35c: Status 404 returned error can't find the container with id 4b0222056e99f1fd8ff216f9fa96a36f3fac6b2f36d9e21b4386e1d58ecad35c Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.282324 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.282685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-trusted-ca\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.282736 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:15.782705882 +0000 UTC m=+147.771763823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283336 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283481 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ksch\" (UniqueName: \"kubernetes.io/projected/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-kube-api-access-8ksch\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283507 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfjr7\" (UniqueName: \"kubernetes.io/projected/9ad71bbd-aa2a-4637-b894-d0b384822a2f-kube-api-access-tfjr7\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283582 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-secret-volume\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283633 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-tls\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283656 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfa02e3-316f-497a-a7d7-ad6d9315efa6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gjtqv\" (UID: \"1dfa02e3-316f-497a-a7d7-ad6d9315efa6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283719 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/075bdaa2-f2c7-4c5a-b0ac-09da9c149141-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dmlw7\" (UID: \"075bdaa2-f2c7-4c5a-b0ac-09da9c149141\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283758 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41e0738b-fa8f-425b-8657-452c59c0b8fc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlgb8\" (UniqueName: \"kubernetes.io/projected/b242c027-08d1-4637-9c6a-0b39f07bcf1f-kube-api-access-dlgb8\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283801 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shnmv\" (UniqueName: \"kubernetes.io/projected/1dfa02e3-316f-497a-a7d7-ad6d9315efa6-kube-api-access-shnmv\") pod \"package-server-manager-789f6589d5-gjtqv\" (UID: \"1dfa02e3-316f-497a-a7d7-ad6d9315efa6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5516996f-81ea-44d0-b5a1-5af61f00af8c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a270ca67-780d-4063-9618-8907cd296ea3-certs\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283868 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b58c6b-55cb-48fd-bdca-1cf449efa230-config\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283915 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s8nz\" (UniqueName: \"kubernetes.io/projected/075bdaa2-f2c7-4c5a-b0ac-09da9c149141-kube-api-access-7s8nz\") pod \"control-plane-machine-set-operator-78cbb6b69f-dmlw7\" (UID: \"075bdaa2-f2c7-4c5a-b0ac-09da9c149141\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.283936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284011 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgqft\" (UniqueName: \"kubernetes.io/projected/33932716-6af1-411f-88c9-a63a12b4077a-kube-api-access-qgqft\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284034 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b84q4\" (UniqueName: \"kubernetes.io/projected/17390854-a3b3-4906-b28b-5a84a5064964-kube-api-access-b84q4\") pod \"multus-admission-controller-857f4d67dd-rhrgq\" (UID: \"17390854-a3b3-4906-b28b-5a84a5064964\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284074 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d69f65ac-9803-46d7-a064-4ee35ffa1877-config-volume\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww45l\" (UniqueName: \"kubernetes.io/projected/a270ca67-780d-4063-9618-8907cd296ea3-kube-api-access-ww45l\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk82d\" (UniqueName: \"kubernetes.io/projected/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-kube-api-access-nk82d\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284491 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t85fp\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-kube-api-access-t85fp\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-serving-cert\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284548 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggkw\" (UniqueName: \"kubernetes.io/projected/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-kube-api-access-2ggkw\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284573 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cea4783-96b7-41bd-8a1f-3893008364a2-proxy-tls\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284608 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/41e0738b-fa8f-425b-8657-452c59c0b8fc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284630 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9afd374c-6bde-459d-9c6c-61d7ed9cad95-metrics-tls\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284699 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9afd374c-6bde-459d-9c6c-61d7ed9cad95-trusted-ca\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-webhook-cert\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284772 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-trusted-ca\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284793 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.284830 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bwbt\" (UniqueName: \"kubernetes.io/projected/41e0738b-fa8f-425b-8657-452c59c0b8fc-kube-api-access-9bwbt\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.291973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.292546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/33932716-6af1-411f-88c9-a63a12b4077a-audit-dir\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.292609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9afd374c-6bde-459d-9c6c-61d7ed9cad95-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.292888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnmdt\" (UniqueName: \"kubernetes.io/projected/c0003968-eec0-455a-ae75-e57629ff675d-kube-api-access-tnmdt\") pod \"migrator-59844c95c7-gc7mr\" (UID: \"c0003968-eec0-455a-ae75-e57629ff675d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.293073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a270ca67-780d-4063-9618-8907cd296ea3-node-bootstrap-token\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.293522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d69f65ac-9803-46d7-a064-4ee35ffa1877-metrics-tls\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.294576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17390854-a3b3-4906-b28b-5a84a5064964-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rhrgq\" (UID: \"17390854-a3b3-4906-b28b-5a84a5064964\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.294653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1cea4783-96b7-41bd-8a1f-3893008364a2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.297038 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41e0738b-fa8f-425b-8657-452c59c0b8fc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.297102 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.297166 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-stats-auth\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.297808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-service-ca-bundle\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.297842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdw7c\" (UniqueName: \"kubernetes.io/projected/1cea4783-96b7-41bd-8a1f-3893008364a2-kube-api-access-rdw7c\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.298047 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64vtz\" (UniqueName: \"kubernetes.io/projected/b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e-kube-api-access-64vtz\") pod \"ingress-canary-wvscw\" (UID: \"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e\") " pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.298073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-tmpfs\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.298595 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90b58c6b-55cb-48fd-bdca-1cf449efa230-serving-cert\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.298629 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kx6s\" (UniqueName: \"kubernetes.io/projected/d69f65ac-9803-46d7-a064-4ee35ffa1877-kube-api-access-2kx6s\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.298715 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-srv-cert\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.299014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnhhk\" (UniqueName: \"kubernetes.io/projected/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-kube-api-access-qnhhk\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.301276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-etcd-client\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.301314 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-registration-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.301569 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e-cert\") pod \"ingress-canary-wvscw\" (UID: \"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e\") " pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.301915 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/66a72013-6e6c-4380-9f5a-c306156e4705-signing-key\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.301945 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cea4783-96b7-41bd-8a1f-3893008364a2-images\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.301971 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/66a72013-6e6c-4380-9f5a-c306156e4705-signing-cabundle\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.301990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-plugins-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.302014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-kube-api-access-5n6zx\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.302145 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-csi-data-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.302466 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-encryption-config\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.302491 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b242c027-08d1-4637-9c6a-0b39f07bcf1f-proxy-tls\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.302788 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/41e0738b-fa8f-425b-8657-452c59c0b8fc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.304201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-config-volume\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5516996f-81ea-44d0-b5a1-5af61f00af8c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307394 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-default-certificate\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307448 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-apiservice-cert\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307747 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-audit-policies\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307809 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-metrics-certs\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307827 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-socket-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.307849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95fx7\" (UniqueName: \"kubernetes.io/projected/9afd374c-6bde-459d-9c6c-61d7ed9cad95-kube-api-access-95fx7\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308650 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-certificates\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.308676 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:15.808656686 +0000 UTC m=+147.797714607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308740 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9ad71bbd-aa2a-4637-b894-d0b384822a2f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-mountpoint-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308827 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8g2b\" (UniqueName: \"kubernetes.io/projected/90b58c6b-55cb-48fd-bdca-1cf449efa230-kube-api-access-b8g2b\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-bound-sa-token\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308881 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8h88\" (UniqueName: \"kubernetes.io/projected/66a72013-6e6c-4380-9f5a-c306156e4705-kube-api-access-z8h88\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5516996f-81ea-44d0-b5a1-5af61f00af8c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8lds\" (UniqueName: \"kubernetes.io/projected/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-kube-api-access-f8lds\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsggw\" (UniqueName: \"kubernetes.io/projected/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-kube-api-access-zsggw\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308980 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vnrm\" (UniqueName: \"kubernetes.io/projected/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-kube-api-access-6vnrm\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.308998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz5n7\" (UniqueName: \"kubernetes.io/projected/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-kube-api-access-zz5n7\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.309029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-srv-cert\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.309264 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b242c027-08d1-4637-9c6a-0b39f07bcf1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.309725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad71bbd-aa2a-4637-b894-d0b384822a2f-serving-cert\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.309923 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41e0738b-fa8f-425b-8657-452c59c0b8fc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.310042 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.310112 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5516996f-81ea-44d0-b5a1-5af61f00af8c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.310380 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-certificates\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.315812 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.315939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-tls\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.331563 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t85fp\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-kube-api-access-t85fp\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.360007 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bwbt\" (UniqueName: \"kubernetes.io/projected/41e0738b-fa8f-425b-8657-452c59c0b8fc-kube-api-access-9bwbt\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.382267 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnhhk\" (UniqueName: \"kubernetes.io/projected/997ec7cb-5a17-4bd9-8da0-774811ee2ac0-kube-api-access-qnhhk\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4hpd\" (UID: \"997ec7cb-5a17-4bd9-8da0-774811ee2ac0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.400345 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-bound-sa-token\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411290 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-kube-api-access-5n6zx\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411513 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-csi-data-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-encryption-config\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411546 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b242c027-08d1-4637-9c6a-0b39f07bcf1f-proxy-tls\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-config-volume\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411580 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-default-certificate\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411595 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-apiservice-cert\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-audit-policies\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-metrics-certs\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-socket-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411677 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95fx7\" (UniqueName: \"kubernetes.io/projected/9afd374c-6bde-459d-9c6c-61d7ed9cad95-kube-api-access-95fx7\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411695 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411711 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9ad71bbd-aa2a-4637-b894-d0b384822a2f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411726 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-mountpoint-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8g2b\" (UniqueName: \"kubernetes.io/projected/90b58c6b-55cb-48fd-bdca-1cf449efa230-kube-api-access-b8g2b\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411758 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8h88\" (UniqueName: \"kubernetes.io/projected/66a72013-6e6c-4380-9f5a-c306156e4705-kube-api-access-z8h88\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8lds\" (UniqueName: \"kubernetes.io/projected/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-kube-api-access-f8lds\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.411799 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:15.911776186 +0000 UTC m=+147.900834097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsggw\" (UniqueName: \"kubernetes.io/projected/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-kube-api-access-zsggw\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411871 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vnrm\" (UniqueName: \"kubernetes.io/projected/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-kube-api-access-6vnrm\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz5n7\" (UniqueName: \"kubernetes.io/projected/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-kube-api-access-zz5n7\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411909 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-srv-cert\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b242c027-08d1-4637-9c6a-0b39f07bcf1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411943 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad71bbd-aa2a-4637-b894-d0b384822a2f-serving-cert\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.411956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-csi-data-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412033 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ksch\" (UniqueName: \"kubernetes.io/projected/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-kube-api-access-8ksch\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412062 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfjr7\" (UniqueName: \"kubernetes.io/projected/9ad71bbd-aa2a-4637-b894-d0b384822a2f-kube-api-access-tfjr7\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-secret-volume\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412123 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/075bdaa2-f2c7-4c5a-b0ac-09da9c149141-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dmlw7\" (UID: \"075bdaa2-f2c7-4c5a-b0ac-09da9c149141\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412161 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfa02e3-316f-497a-a7d7-ad6d9315efa6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gjtqv\" (UID: \"1dfa02e3-316f-497a-a7d7-ad6d9315efa6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412179 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlgb8\" (UniqueName: \"kubernetes.io/projected/b242c027-08d1-4637-9c6a-0b39f07bcf1f-kube-api-access-dlgb8\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412194 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shnmv\" (UniqueName: \"kubernetes.io/projected/1dfa02e3-316f-497a-a7d7-ad6d9315efa6-kube-api-access-shnmv\") pod \"package-server-manager-789f6589d5-gjtqv\" (UID: \"1dfa02e3-316f-497a-a7d7-ad6d9315efa6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a270ca67-780d-4063-9618-8907cd296ea3-certs\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b58c6b-55cb-48fd-bdca-1cf449efa230-config\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s8nz\" (UniqueName: \"kubernetes.io/projected/075bdaa2-f2c7-4c5a-b0ac-09da9c149141-kube-api-access-7s8nz\") pod \"control-plane-machine-set-operator-78cbb6b69f-dmlw7\" (UID: \"075bdaa2-f2c7-4c5a-b0ac-09da9c149141\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgqft\" (UniqueName: \"kubernetes.io/projected/33932716-6af1-411f-88c9-a63a12b4077a-kube-api-access-qgqft\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412280 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b84q4\" (UniqueName: \"kubernetes.io/projected/17390854-a3b3-4906-b28b-5a84a5064964-kube-api-access-b84q4\") pod \"multus-admission-controller-857f4d67dd-rhrgq\" (UID: \"17390854-a3b3-4906-b28b-5a84a5064964\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412328 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d69f65ac-9803-46d7-a064-4ee35ffa1877-config-volume\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412363 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww45l\" (UniqueName: \"kubernetes.io/projected/a270ca67-780d-4063-9618-8907cd296ea3-kube-api-access-ww45l\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk82d\" (UniqueName: \"kubernetes.io/projected/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-kube-api-access-nk82d\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412398 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-serving-cert\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggkw\" (UniqueName: \"kubernetes.io/projected/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-kube-api-access-2ggkw\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cea4783-96b7-41bd-8a1f-3893008364a2-proxy-tls\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412466 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9afd374c-6bde-459d-9c6c-61d7ed9cad95-metrics-tls\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412486 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9afd374c-6bde-459d-9c6c-61d7ed9cad95-trusted-ca\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412502 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-webhook-cert\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412539 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/33932716-6af1-411f-88c9-a63a12b4077a-audit-dir\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9afd374c-6bde-459d-9c6c-61d7ed9cad95-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412574 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnmdt\" (UniqueName: \"kubernetes.io/projected/c0003968-eec0-455a-ae75-e57629ff675d-kube-api-access-tnmdt\") pod \"migrator-59844c95c7-gc7mr\" (UID: \"c0003968-eec0-455a-ae75-e57629ff675d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412590 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a270ca67-780d-4063-9618-8907cd296ea3-node-bootstrap-token\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412607 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d69f65ac-9803-46d7-a064-4ee35ffa1877-metrics-tls\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17390854-a3b3-4906-b28b-5a84a5064964-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rhrgq\" (UID: \"17390854-a3b3-4906-b28b-5a84a5064964\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412643 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1cea4783-96b7-41bd-8a1f-3893008364a2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412661 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412680 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-stats-auth\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-service-ca-bundle\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412714 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdw7c\" (UniqueName: \"kubernetes.io/projected/1cea4783-96b7-41bd-8a1f-3893008364a2-kube-api-access-rdw7c\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412730 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-tmpfs\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64vtz\" (UniqueName: \"kubernetes.io/projected/b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e-kube-api-access-64vtz\") pod \"ingress-canary-wvscw\" (UID: \"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e\") " pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412765 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90b58c6b-55cb-48fd-bdca-1cf449efa230-serving-cert\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412781 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kx6s\" (UniqueName: \"kubernetes.io/projected/d69f65ac-9803-46d7-a064-4ee35ffa1877-kube-api-access-2kx6s\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412797 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-srv-cert\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412818 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-etcd-client\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-registration-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412847 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e-cert\") pod \"ingress-canary-wvscw\" (UID: \"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e\") " pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412865 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/66a72013-6e6c-4380-9f5a-c306156e4705-signing-key\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cea4783-96b7-41bd-8a1f-3893008364a2-images\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412904 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/66a72013-6e6c-4380-9f5a-c306156e4705-signing-cabundle\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.412920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-plugins-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.413161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-plugins-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.415538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-encryption-config\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.419719 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-service-ca-bundle\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.421041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9afd374c-6bde-459d-9c6c-61d7ed9cad95-trusted-ca\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.427087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-config-volume\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.427272 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b242c027-08d1-4637-9c6a-0b39f07bcf1f-proxy-tls\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.427710 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/33932716-6af1-411f-88c9-a63a12b4077a-audit-dir\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.428267 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b58c6b-55cb-48fd-bdca-1cf449efa230-config\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.428640 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1cea4783-96b7-41bd-8a1f-3893008364a2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.428967 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.429724 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.430433 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d69f65ac-9803-46d7-a064-4ee35ffa1877-config-volume\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.432229 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b242c027-08d1-4637-9c6a-0b39f07bcf1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.434593 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.437641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a270ca67-780d-4063-9618-8907cd296ea3-certs\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.439554 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ad71bbd-aa2a-4637-b894-d0b384822a2f-serving-cert\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.439553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-srv-cert\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.439904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.440103 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/66a72013-6e6c-4380-9f5a-c306156e4705-signing-key\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.440784 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-tmpfs\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.441770 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1cea4783-96b7-41bd-8a1f-3893008364a2-images\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.443384 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-default-certificate\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.446533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1cea4783-96b7-41bd-8a1f-3893008364a2-proxy-tls\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.446863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d69f65ac-9803-46d7-a064-4ee35ffa1877-metrics-tls\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.447336 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9afd374c-6bde-459d-9c6c-61d7ed9cad95-metrics-tls\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.447492 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/17390854-a3b3-4906-b28b-5a84a5064964-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rhrgq\" (UID: \"17390854-a3b3-4906-b28b-5a84a5064964\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.447910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-stats-auth\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.448223 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.448225 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9ad71bbd-aa2a-4637-b894-d0b384822a2f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.448827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/33932716-6af1-411f-88c9-a63a12b4077a-audit-policies\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.449014 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-socket-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.451450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-apiservice-cert\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.453373 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfa02e3-316f-497a-a7d7-ad6d9315efa6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gjtqv\" (UID: \"1dfa02e3-316f-497a-a7d7-ad6d9315efa6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.454258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.455135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-mountpoint-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.455330 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-registration-dir\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.455537 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.456100 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-etcd-client\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.456474 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.462567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33932716-6af1-411f-88c9-a63a12b4077a-serving-cert\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.462680 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90b58c6b-55cb-48fd-bdca-1cf449efa230-serving-cert\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.463307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/66a72013-6e6c-4380-9f5a-c306156e4705-signing-cabundle\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.467246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.471179 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-metrics-certs\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.471641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.473808 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/075bdaa2-f2c7-4c5a-b0ac-09da9c149141-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dmlw7\" (UID: \"075bdaa2-f2c7-4c5a-b0ac-09da9c149141\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.478270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a270ca67-780d-4063-9618-8907cd296ea3-node-bootstrap-token\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.481192 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-srv-cert\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.484809 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-webhook-cert\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.487587 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-kube-api-access-5n6zx\") pod \"marketplace-operator-79b997595-qpjtj\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.490689 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-secret-volume\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.491982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e-cert\") pod \"ingress-canary-wvscw\" (UID: \"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e\") " pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.492562 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41e0738b-fa8f-425b-8657-452c59c0b8fc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-72sx7\" (UID: \"41e0738b-fa8f-425b-8657-452c59c0b8fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.493180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8lds\" (UniqueName: \"kubernetes.io/projected/a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a-kube-api-access-f8lds\") pod \"olm-operator-6b444d44fb-xstng\" (UID: \"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.503987 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsggw\" (UniqueName: \"kubernetes.io/projected/c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b-kube-api-access-zsggw\") pod \"packageserver-d55dfcdfc-p5bxp\" (UID: \"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.514166 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.514704 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.014691949 +0000 UTC m=+148.003749870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.529026 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vnrm\" (UniqueName: \"kubernetes.io/projected/7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1-kube-api-access-6vnrm\") pod \"csi-hostpathplugin-wr5kd\" (UID: \"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1\") " pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.541614 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-955gc"] Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.548614 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz5n7\" (UniqueName: \"kubernetes.io/projected/21c2490b-9fde-43ae-ab99-8b5b3758fdd5-kube-api-access-zz5n7\") pod \"kube-storage-version-migrator-operator-b67b599dd-swh9k\" (UID: \"21c2490b-9fde-43ae-ab99-8b5b3758fdd5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.572079 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.578469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s8nz\" (UniqueName: \"kubernetes.io/projected/075bdaa2-f2c7-4c5a-b0ac-09da9c149141-kube-api-access-7s8nz\") pod \"control-plane-machine-set-operator-78cbb6b69f-dmlw7\" (UID: \"075bdaa2-f2c7-4c5a-b0ac-09da9c149141\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.583788 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.594133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgqft\" (UniqueName: \"kubernetes.io/projected/33932716-6af1-411f-88c9-a63a12b4077a-kube-api-access-qgqft\") pod \"apiserver-7bbb656c7d-7m2jp\" (UID: \"33932716-6af1-411f-88c9-a63a12b4077a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.596926 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nmgd9"] Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.600218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b84q4\" (UniqueName: \"kubernetes.io/projected/17390854-a3b3-4906-b28b-5a84a5064964-kube-api-access-b84q4\") pod \"multus-admission-controller-857f4d67dd-rhrgq\" (UID: \"17390854-a3b3-4906-b28b-5a84a5064964\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.623708 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.624575 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.624934 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.124919787 +0000 UTC m=+148.113977708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.631359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.645035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww45l\" (UniqueName: \"kubernetes.io/projected/a270ca67-780d-4063-9618-8907cd296ea3-kube-api-access-ww45l\") pod \"machine-config-server-lj9cj\" (UID: \"a270ca67-780d-4063-9618-8907cd296ea3\") " pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.655257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk82d\" (UniqueName: \"kubernetes.io/projected/85f3bf48-28bd-4afe-b0d9-a0fbd11de675-kube-api-access-nk82d\") pod \"router-default-5444994796-q6bv4\" (UID: \"85f3bf48-28bd-4afe-b0d9-a0fbd11de675\") " pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.656109 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.657127 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9afd374c-6bde-459d-9c6c-61d7ed9cad95-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.678716 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:15 crc kubenswrapper[4744]: W1201 08:20:15.685028 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod214a177a_1cc2_4322_8994_400d072af8e7.slice/crio-34e4b89971823adec5abcbb38f172a1db522fa52ec9d95ce578de5d825602ec1 WatchSource:0}: Error finding container 34e4b89971823adec5abcbb38f172a1db522fa52ec9d95ce578de5d825602ec1: Status 404 returned error can't find the container with id 34e4b89971823adec5abcbb38f172a1db522fa52ec9d95ce578de5d825602ec1 Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.698364 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnmdt\" (UniqueName: \"kubernetes.io/projected/c0003968-eec0-455a-ae75-e57629ff675d-kube-api-access-tnmdt\") pod \"migrator-59844c95c7-gc7mr\" (UID: \"c0003968-eec0-455a-ae75-e57629ff675d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.699399 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s"] Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.699962 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.717643 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6"] Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.721487 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lj9cj" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.731327 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.731799 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.231783281 +0000 UTC m=+148.220841202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.734011 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.739157 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-l9mth"] Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.739932 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64vtz\" (UniqueName: \"kubernetes.io/projected/b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e-kube-api-access-64vtz\") pod \"ingress-canary-wvscw\" (UID: \"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e\") " pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.743967 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdw7c\" (UniqueName: \"kubernetes.io/projected/1cea4783-96b7-41bd-8a1f-3893008364a2-kube-api-access-rdw7c\") pod \"machine-config-operator-74547568cd-8jr2s\" (UID: \"1cea4783-96b7-41bd-8a1f-3893008364a2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.750250 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z"] Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.765822 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8g2b\" (UniqueName: \"kubernetes.io/projected/90b58c6b-55cb-48fd-bdca-1cf449efa230-kube-api-access-b8g2b\") pod \"service-ca-operator-777779d784-nl48d\" (UID: \"90b58c6b-55cb-48fd-bdca-1cf449efa230\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.779759 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggkw\" (UniqueName: \"kubernetes.io/projected/f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd-kube-api-access-2ggkw\") pod \"catalog-operator-68c6474976-2ghv7\" (UID: \"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.780097 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.796306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlgb8\" (UniqueName: \"kubernetes.io/projected/b242c027-08d1-4637-9c6a-0b39f07bcf1f-kube-api-access-dlgb8\") pod \"machine-config-controller-84d6567774-7th4n\" (UID: \"b242c027-08d1-4637-9c6a-0b39f07bcf1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.814578 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shnmv\" (UniqueName: \"kubernetes.io/projected/1dfa02e3-316f-497a-a7d7-ad6d9315efa6-kube-api-access-shnmv\") pod \"package-server-manager-789f6589d5-gjtqv\" (UID: \"1dfa02e3-316f-497a-a7d7-ad6d9315efa6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.832930 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.833536 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.333395948 +0000 UTC m=+148.322453869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.833769 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.835072 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.335041374 +0000 UTC m=+148.324099295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.844108 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ksch\" (UniqueName: \"kubernetes.io/projected/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-kube-api-access-8ksch\") pod \"collect-profiles-29409615-l7t55\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.851246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.862118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfjr7\" (UniqueName: \"kubernetes.io/projected/9ad71bbd-aa2a-4637-b894-d0b384822a2f-kube-api-access-tfjr7\") pod \"openshift-config-operator-7777fb866f-44bgh\" (UID: \"9ad71bbd-aa2a-4637-b894-d0b384822a2f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.875511 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.890675 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.890950 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95fx7\" (UniqueName: \"kubernetes.io/projected/9afd374c-6bde-459d-9c6c-61d7ed9cad95-kube-api-access-95fx7\") pod \"ingress-operator-5b745b69d9-fh8z6\" (UID: \"9afd374c-6bde-459d-9c6c-61d7ed9cad95\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.908074 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.917789 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kx6s\" (UniqueName: \"kubernetes.io/projected/d69f65ac-9803-46d7-a064-4ee35ffa1877-kube-api-access-2kx6s\") pod \"dns-default-ml6wn\" (UID: \"d69f65ac-9803-46d7-a064-4ee35ffa1877\") " pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.920117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8h88\" (UniqueName: \"kubernetes.io/projected/66a72013-6e6c-4380-9f5a-c306156e4705-kube-api-access-z8h88\") pod \"service-ca-9c57cc56f-5kq46\" (UID: \"66a72013-6e6c-4380-9f5a-c306156e4705\") " pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.935248 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.935420 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.435383085 +0000 UTC m=+148.424441006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.936757 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:15 crc kubenswrapper[4744]: E1201 08:20:15.937302 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.437289578 +0000 UTC m=+148.426347499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.940234 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" Dec 01 08:20:15 crc kubenswrapper[4744]: W1201 08:20:15.941338 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c6ffc7f_409c_4a41_b53d_38a6c85b5fcc.slice/crio-37fc9c1f3bcace38e114806b6800dc38ce60bed24063d7549270ae2b6d06eb48 WatchSource:0}: Error finding container 37fc9c1f3bcace38e114806b6800dc38ce60bed24063d7549270ae2b6d06eb48: Status 404 returned error can't find the container with id 37fc9c1f3bcace38e114806b6800dc38ce60bed24063d7549270ae2b6d06eb48 Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.945753 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.954121 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.962765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" Dec 01 08:20:15 crc kubenswrapper[4744]: I1201 08:20:15.970314 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.010128 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-wvscw" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.039867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.040448 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.540432908 +0000 UTC m=+148.529490829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.045728 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.059662 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.141121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.141488 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.6414751 +0000 UTC m=+148.630533021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.141845 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.148818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ef6e8553431fccea1fbf750569ce3d5935a5a205d835f048889ee3d84e314fe0"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.148880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8c79b4a0bf49e55a5ef64084c839ea07b66b4bc23d32e159514ba23634466f14"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.151011 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lj9cj" event={"ID":"a270ca67-780d-4063-9618-8907cd296ea3","Type":"ContainerStarted","Data":"7eed20ec78b70272d1f5b87627a2b7e53efb41758bad41c2cd000ab044adaaa9"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.153822 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" event={"ID":"b80c9c0b-07a4-441e-9f71-f9962901967e","Type":"ContainerStarted","Data":"cc42d81a8d018371f8282a534761d6aee6de9921e9f1594ce9f33decc048b9fd"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.156704 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" event={"ID":"268ee28c-0643-4161-aee8-6fa5a08722c8","Type":"ContainerStarted","Data":"df7f6e0302cc19d9f5ec354e76010432c767cdf876af63d755d3e873336a1b27"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.166122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" event={"ID":"214a177a-1cc2-4322-8994-400d072af8e7","Type":"ContainerStarted","Data":"34e4b89971823adec5abcbb38f172a1db522fa52ec9d95ce578de5d825602ec1"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.168345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" event={"ID":"54063efd-02f0-46a6-8f06-edd7070fda31","Type":"ContainerStarted","Data":"d20a33b3a9536d2c9e3c6edc120215aadcd27d78e1469ddd024884788a995877"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.175493 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nmgd9" event={"ID":"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10","Type":"ContainerStarted","Data":"5dd76aafec99c08411ad2d2643b8b5bb842f980f8b571709e8702333e4b267c2"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.186283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" event={"ID":"9a9c9a23-151c-41ce-9a1b-57adbd823c1e","Type":"ContainerStarted","Data":"4b0222056e99f1fd8ff216f9fa96a36f3fac6b2f36d9e21b4386e1d58ecad35c"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.197657 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"dabf54179f4162b19115dc3d2da111fead308bb3245d720d9d237b389df6a9a0"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.197731 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a701a286b2f661f6b89d7466cbb1fa9cd0903a59fc6268723d0efb269efdabf9"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.198040 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.212599 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" event={"ID":"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc","Type":"ContainerStarted","Data":"37fc9c1f3bcace38e114806b6800dc38ce60bed24063d7549270ae2b6d06eb48"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.222133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" event={"ID":"41c7aa1f-b684-4370-ab59-87aeb78251cd","Type":"ContainerStarted","Data":"965ef4e43088e953348afdcf906d15a58663514bfc38b42b469ef1e9102392ec"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.223712 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4a979e394978a9f79204e661e066ccd4b992382f7c47fd55a9614c72ba21f30f"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.223758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a0acc4e47f7400679754462b93f3a43cd349bf05099f6e3e11387bec177557dd"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.227334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dk5mh" event={"ID":"a368aebd-97fc-44cf-bd2b-755850a3b5d2","Type":"ContainerStarted","Data":"506ffffdf0a65c5aa5ebe1ca178e0310ee6b6674bb103c0fd153052ea57c5539"} Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.227707 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.232388 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-dk5mh" Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.241954 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.243288 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.743272752 +0000 UTC m=+148.732330673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: W1201 08:20:16.264184 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85f3bf48_28bd_4afe_b0d9_a0fbd11de675.slice/crio-cf3a09e23e955e240ca9f429e209980258c91907b548555476d6602b9e887a2b WatchSource:0}: Error finding container cf3a09e23e955e240ca9f429e209980258c91907b548555476d6602b9e887a2b: Status 404 returned error can't find the container with id cf3a09e23e955e240ca9f429e209980258c91907b548555476d6602b9e887a2b Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.352028 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd"] Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.353750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.354555 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.854540059 +0000 UTC m=+148.843597980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.369447 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg"] Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.373154 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6qsl"] Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.463149 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.463676 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:16.963640565 +0000 UTC m=+148.952698486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.569194 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.571593 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.071575979 +0000 UTC m=+149.060633900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.671959 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.672295 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.17227958 +0000 UTC m=+149.161337501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.777485 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.779999 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.279980257 +0000 UTC m=+149.269038178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.860275 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l"] Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.860324 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pflgv"] Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.877806 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mvtcw"] Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.886067 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mq727"] Dec 01 08:20:16 crc kubenswrapper[4744]: I1201 08:20:16.903901 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:16 crc kubenswrapper[4744]: E1201 08:20:16.904435 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.404395211 +0000 UTC m=+149.393453132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.006084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.006750 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.506738589 +0000 UTC m=+149.495796510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.054609 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-dk5mh" podStartSLOduration=124.054594635 podStartE2EDuration="2m4.054594635s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.025885323 +0000 UTC m=+149.014943244" watchObservedRunningTime="2025-12-01 08:20:17.054594635 +0000 UTC m=+149.043652556" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.107674 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.107886 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.607862982 +0000 UTC m=+149.596920903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.108529 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.109582 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.60956654 +0000 UTC m=+149.598624461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.221515 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.221632 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.721608188 +0000 UTC m=+149.710666109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.221914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.222302 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.722295537 +0000 UTC m=+149.711353458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.243801 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.261214 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" event={"ID":"b80c9c0b-07a4-441e-9f71-f9962901967e","Type":"ContainerStarted","Data":"e4c5113cd6dbcebcac9766fb39eb781b3354b10aa2c1fe26df0f6ee7d275bfea"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.264341 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m"] Dec 01 08:20:17 crc kubenswrapper[4744]: W1201 08:20:17.294712 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30a4c368_98eb_4f52_be7b_0d6aa7d50a55.slice/crio-3c9c23776479485ddf73a499e9d63347e107410e5d0f4f3b22fa6277dece4d5a WatchSource:0}: Error finding container 3c9c23776479485ddf73a499e9d63347e107410e5d0f4f3b22fa6277dece4d5a: Status 404 returned error can't find the container with id 3c9c23776479485ddf73a499e9d63347e107410e5d0f4f3b22fa6277dece4d5a Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.295942 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" event={"ID":"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc","Type":"ContainerStarted","Data":"d7890b7b8e3ea69b060e1088c86c6e6198046ecd5db0a553f42d50f1755eb6d5"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.298876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" event={"ID":"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61","Type":"ContainerStarted","Data":"332b06aaabb88f879f1d94955a195195faf99bc8de6f62095b6aad17e46a3a1f"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.307932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" event={"ID":"f71e09c5-239a-4982-ab7d-93c6d647048a","Type":"ContainerStarted","Data":"6b517d218f0a3f461fb6d739e7a3611104f6dab472569a18970b1fca36411f6c"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.313170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-q6bv4" event={"ID":"85f3bf48-28bd-4afe-b0d9-a0fbd11de675","Type":"ContainerStarted","Data":"3577ccbbb918d228c15b304f9246994ae70b7e2e21a21d7b5474522834a21c33"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.313213 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-q6bv4" event={"ID":"85f3bf48-28bd-4afe-b0d9-a0fbd11de675","Type":"ContainerStarted","Data":"cf3a09e23e955e240ca9f429e209980258c91907b548555476d6602b9e887a2b"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.319843 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pnhzx"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.320588 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nmgd9" event={"ID":"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10","Type":"ContainerStarted","Data":"c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.323045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.323527 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.823501143 +0000 UTC m=+149.812559064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.323599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.324131 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.82411406 +0000 UTC m=+149.813171981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.325317 4744 generic.go:334] "Generic (PLEG): container finished" podID="9a9c9a23-151c-41ce-9a1b-57adbd823c1e" containerID="987e8bc096db0e87e7399da500bb96458cdbd42d39472fb428c629269d9e2d85" exitCode=0 Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.325490 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" event={"ID":"9a9c9a23-151c-41ce-9a1b-57adbd823c1e","Type":"ContainerDied","Data":"987e8bc096db0e87e7399da500bb96458cdbd42d39472fb428c629269d9e2d85"} Dec 01 08:20:17 crc kubenswrapper[4744]: W1201 08:20:17.336259 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod948afa07_87fa_4c8d_a26d_5b9b8d3b156a.slice/crio-de8ebdd6b6c3c4b8aa46fafa2efde83ffb9b981e91b0b075aa62ca997016b8b3 WatchSource:0}: Error finding container de8ebdd6b6c3c4b8aa46fafa2efde83ffb9b981e91b0b075aa62ca997016b8b3: Status 404 returned error can't find the container with id de8ebdd6b6c3c4b8aa46fafa2efde83ffb9b981e91b0b075aa62ca997016b8b3 Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.369390 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qpjtj"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.369701 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lh52z" podStartSLOduration=124.369687683 podStartE2EDuration="2m4.369687683s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.339239693 +0000 UTC m=+149.328297614" watchObservedRunningTime="2025-12-01 08:20:17.369687683 +0000 UTC m=+149.358745604" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.370440 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" event={"ID":"214a177a-1cc2-4322-8994-400d072af8e7","Type":"ContainerStarted","Data":"80c58d71d4e0307214afb9c16d2b68eae1dd57602ef0cd372531823184e2c835"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.375515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" event={"ID":"f71710a5-e0fa-4ac7-be6a-80e84a342e3b","Type":"ContainerStarted","Data":"e3040f07b5660a556747e1d02ea33f0b4a1e0e3f73cc7004a34a5e7b42bd2e2a"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.377202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" event={"ID":"a87bf7b3-7b9d-407d-a113-387afac58cab","Type":"ContainerStarted","Data":"c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.377222 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" event={"ID":"a87bf7b3-7b9d-407d-a113-387afac58cab","Type":"ContainerStarted","Data":"fc2d0e363778c0ee556e8c3b9d1ffafb4781b225f5ac4f139ee908064342096b"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.377944 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.379499 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" event={"ID":"54063efd-02f0-46a6-8f06-edd7070fda31","Type":"ContainerStarted","Data":"eb4e01edc14cf7b14a2208ef0865493b868f8f84a54aac5f4e867ab1e5afad60"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.396686 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-nmgd9" podStartSLOduration=124.396666026 podStartE2EDuration="2m4.396666026s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.393832217 +0000 UTC m=+149.382890158" watchObservedRunningTime="2025-12-01 08:20:17.396666026 +0000 UTC m=+149.385723947" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.398262 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.398311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" event={"ID":"41c7aa1f-b684-4370-ab59-87aeb78251cd","Type":"ContainerStarted","Data":"d15b506814c72bffff364fe6f2471fdf49a2f34ded19b0d172c05d6237d5fc9d"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.400631 4744 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-r6qsl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.400678 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" podUID="a87bf7b3-7b9d-407d-a113-387afac58cab" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.403301 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" event={"ID":"ce21f093-ef49-49f7-b8aa-e1159c94b30e","Type":"ContainerStarted","Data":"f5592d41593bbbfea5320b79823b3862add62cd1510981be6286f05c38e4edc3"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.403339 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" event={"ID":"ce21f093-ef49-49f7-b8aa-e1159c94b30e","Type":"ContainerStarted","Data":"26f712f009d8009d65103f38ee9d7cdc01032272ffb9ed7fba06e59172269d03"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.424462 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.424734 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.924717379 +0000 UTC m=+149.913775290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.425184 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lj9cj" event={"ID":"a270ca67-780d-4063-9618-8907cd296ea3","Type":"ContainerStarted","Data":"7e2ad6d0619989d0af7565c1bbdf4134d8859d5df17a5357e3ceae566584e7a2"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.425338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.441160 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:17.941133778 +0000 UTC m=+149.930191699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.472266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" event={"ID":"268ee28c-0643-4161-aee8-6fa5a08722c8","Type":"ContainerStarted","Data":"c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.473010 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.473065 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-q6bv4" podStartSLOduration=124.473040048 podStartE2EDuration="2m4.473040048s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.470197139 +0000 UTC m=+149.459255060" watchObservedRunningTime="2025-12-01 08:20:17.473040048 +0000 UTC m=+149.462097969" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.504084 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.505024 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" podStartSLOduration=124.505007161 podStartE2EDuration="2m4.505007161s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.503686174 +0000 UTC m=+149.492744095" watchObservedRunningTime="2025-12-01 08:20:17.505007161 +0000 UTC m=+149.494065082" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.520784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" event={"ID":"997ec7cb-5a17-4bd9-8da0-774811ee2ac0","Type":"ContainerStarted","Data":"76be7c32aa87c14ce2d749a46775cc9889287cefd23ab8c7d844a91317595608"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.520833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" event={"ID":"997ec7cb-5a17-4bd9-8da0-774811ee2ac0","Type":"ContainerStarted","Data":"2a1a95cc789f0a71113b66911f4a67496354cf0eab895657cd57aac2aae0dd97"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.531326 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.533294 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.03327776 +0000 UTC m=+150.022335681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.552512 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" podStartSLOduration=124.552494007 podStartE2EDuration="2m4.552494007s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.543153076 +0000 UTC m=+149.532210997" watchObservedRunningTime="2025-12-01 08:20:17.552494007 +0000 UTC m=+149.541551918" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.569861 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.577859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mq727" event={"ID":"5b7ea6bf-6f2c-483e-b37b-9032d56018ee","Type":"ContainerStarted","Data":"4056353692e147b62a40f60cb0a6c1d23e91987926ad3a5270c8a2748598f78c"} Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.577893 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mq727" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.586927 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-mq727 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.586971 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mq727" podUID="5b7ea6bf-6f2c-483e-b37b-9032d56018ee" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.638596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.639803 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.139791284 +0000 UTC m=+150.128849205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.640855 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" podStartSLOduration=124.640840674 podStartE2EDuration="2m4.640840674s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.602890004 +0000 UTC m=+149.591947925" watchObservedRunningTime="2025-12-01 08:20:17.640840674 +0000 UTC m=+149.629898595" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.642142 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzlbv" podStartSLOduration=124.6421369 podStartE2EDuration="2m4.6421369s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.637935222 +0000 UTC m=+149.626993143" watchObservedRunningTime="2025-12-01 08:20:17.6421369 +0000 UTC m=+149.631194821" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.657105 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.737422 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" podStartSLOduration=124.73739666 podStartE2EDuration="2m4.73739666s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.720182379 +0000 UTC m=+149.709240300" watchObservedRunningTime="2025-12-01 08:20:17.73739666 +0000 UTC m=+149.726454581" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.744925 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.745189 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.245174637 +0000 UTC m=+150.234232558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.770213 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-lj9cj" podStartSLOduration=5.770197945 podStartE2EDuration="5.770197945s" podCreationTimestamp="2025-12-01 08:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.767744207 +0000 UTC m=+149.756802128" watchObservedRunningTime="2025-12-01 08:20:17.770197945 +0000 UTC m=+149.759255866" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.791556 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.836772 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-955gc" podStartSLOduration=124.836758794 podStartE2EDuration="2m4.836758794s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.833977386 +0000 UTC m=+149.823035307" watchObservedRunningTime="2025-12-01 08:20:17.836758794 +0000 UTC m=+149.825816715" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.837071 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.851353 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wr5kd"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.851428 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.852213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.852596 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.352581245 +0000 UTC m=+150.341639166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.857507 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whgj6" podStartSLOduration=124.857495062 podStartE2EDuration="2m4.857495062s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.856530245 +0000 UTC m=+149.845588166" watchObservedRunningTime="2025-12-01 08:20:17.857495062 +0000 UTC m=+149.846552983" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.876394 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.883649 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:17 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:17 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:17 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.883697 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.893125 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rhrgq"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.903979 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-wvscw"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.908229 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.924483 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.924774 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pw4dg" podStartSLOduration=124.9247526 podStartE2EDuration="2m4.9247526s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.919648807 +0000 UTC m=+149.908706728" watchObservedRunningTime="2025-12-01 08:20:17.9247526 +0000 UTC m=+149.913810521" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.935383 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ml6wn"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.939143 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.950527 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mq727" podStartSLOduration=124.950510149 podStartE2EDuration="2m4.950510149s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.947869445 +0000 UTC m=+149.936927356" watchObservedRunningTime="2025-12-01 08:20:17.950510149 +0000 UTC m=+149.939568060" Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.951692 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55"] Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.953024 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:17 crc kubenswrapper[4744]: E1201 08:20:17.953293 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.453280386 +0000 UTC m=+150.442338307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:17 crc kubenswrapper[4744]: I1201 08:20:17.987620 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4hpd" podStartSLOduration=124.987591724 podStartE2EDuration="2m4.987591724s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:17.986270397 +0000 UTC m=+149.975328318" watchObservedRunningTime="2025-12-01 08:20:17.987591724 +0000 UTC m=+149.976649645" Dec 01 08:20:18 crc kubenswrapper[4744]: W1201 08:20:18.020500 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0003968_eec0_455a_ae75_e57629ff675d.slice/crio-6fa48379ac4070bad6074eba17182bf6a88d49ee752565ecc319931470c56980 WatchSource:0}: Error finding container 6fa48379ac4070bad6074eba17182bf6a88d49ee752565ecc319931470c56980: Status 404 returned error can't find the container with id 6fa48379ac4070bad6074eba17182bf6a88d49ee752565ecc319931470c56980 Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.036813 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5kq46"] Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.066480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.066844 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.566826287 +0000 UTC m=+150.555884208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.089214 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-44bgh"] Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.115220 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6"] Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.121434 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nl48d"] Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.154760 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n"] Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.173120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.173532 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.673517436 +0000 UTC m=+150.662575357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.274234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.274834 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.774823774 +0000 UTC m=+150.763881685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.375328 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.375491 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.875461304 +0000 UTC m=+150.864519225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.375908 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.376394 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.876345499 +0000 UTC m=+150.865403420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.478640 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.479034 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:18.979019606 +0000 UTC m=+150.968077527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.580548 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.580911 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.08089976 +0000 UTC m=+151.069957681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.633480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mq727" event={"ID":"5b7ea6bf-6f2c-483e-b37b-9032d56018ee","Type":"ContainerStarted","Data":"4cafee4e191cf3dd0c08cb3e98c7c712abe7828adfb4a50cb4a91fc61d417fdc"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.634762 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-mq727 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.634789 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mq727" podUID="5b7ea6bf-6f2c-483e-b37b-9032d56018ee" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.639925 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.640071 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.681712 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.683792 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.182325602 +0000 UTC m=+151.171383523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.684407 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" event={"ID":"30a4c368-98eb-4f52-be7b-0d6aa7d50a55","Type":"ContainerStarted","Data":"e3408b39eab9e03ea1d582b53a54e2a3cd7315bc17342de7f4dc553e6b08cd9b"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.684480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" event={"ID":"30a4c368-98eb-4f52-be7b-0d6aa7d50a55","Type":"ContainerStarted","Data":"3c9c23776479485ddf73a499e9d63347e107410e5d0f4f3b22fa6277dece4d5a"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.731702 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" event={"ID":"c0003968-eec0-455a-ae75-e57629ff675d","Type":"ContainerStarted","Data":"6fa48379ac4070bad6074eba17182bf6a88d49ee752565ecc319931470c56980"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.753057 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" event={"ID":"1cea4783-96b7-41bd-8a1f-3893008364a2","Type":"ContainerStarted","Data":"0f698a0b74e659bed15e1b71a2988f4cb0113eaea5230ad264b5c3c716cf2110"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.753094 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" event={"ID":"1cea4783-96b7-41bd-8a1f-3893008364a2","Type":"ContainerStarted","Data":"8923722b3771329836d235a3154eb09e289f1b9840f3d4de9154690e379a1e07"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.773714 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" event={"ID":"66a72013-6e6c-4380-9f5a-c306156e4705","Type":"ContainerStarted","Data":"c3958aa29c4d2c24cfdab0c17425cef3c33e09ee0f8ac2b85d82f043b1b061a2"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.783735 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.784632 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.284615028 +0000 UTC m=+151.273672949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.818975 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" event={"ID":"9a9c9a23-151c-41ce-9a1b-57adbd823c1e","Type":"ContainerStarted","Data":"4e171d892aa8210211d12153b6aedc3ca1137a2af72f1f0c8da27a3393390fd4"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.830932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" event={"ID":"17390854-a3b3-4906-b28b-5a84a5064964","Type":"ContainerStarted","Data":"81b620d406c4a17b190f90579a1260fff50dcc5681366e93b0426b5771d4ae8e"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.844689 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mvtcw" event={"ID":"f71710a5-e0fa-4ac7-be6a-80e84a342e3b","Type":"ContainerStarted","Data":"63a3de420b55b61c6fa4ab0eb618529c18bcb0069156b669bf101a9ea2476ac4"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.885222 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.886350 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.386327698 +0000 UTC m=+151.375385619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.901194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-l9mth" event={"ID":"6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc","Type":"ContainerStarted","Data":"201ce3d2b4b5adc763f88a8662f9d4a1ca4268864e699e5336a52152279730cb"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.915527 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:18 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:18 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:18 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.915613 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.923025 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ml6wn" event={"ID":"d69f65ac-9803-46d7-a064-4ee35ffa1877","Type":"ContainerStarted","Data":"c477542505fa0b5a05a4a463c82af566f0abcfdf85f51a5888cecdcc691b93b2"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.961769 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-wvscw" event={"ID":"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e","Type":"ContainerStarted","Data":"a3bf59822e5d6d39737622272ca0daf1fbc4beaab3fdd38a65ae2e2b46add23b"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.973134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" event={"ID":"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a","Type":"ContainerStarted","Data":"5c3ca4bb315a4d6792fc5780abcbc59589dc26613cbf0fa3ed563a7500801006"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.973214 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" event={"ID":"a438ecdb-2567-4ed2-bf6a-1cbe75f5fa8a","Type":"ContainerStarted","Data":"61ff0be8c40b5e19091b149464a5639901460a595832a42c919cef0e6f0e1ea1"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.973566 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.975707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" event={"ID":"1dfa02e3-316f-497a-a7d7-ad6d9315efa6","Type":"ContainerStarted","Data":"0c5943ef5db92c92759225e596bd4d0480c52937392bfaf2a6684b2e6ade9f57"} Dec 01 08:20:18 crc kubenswrapper[4744]: I1201 08:20:18.987056 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:18 crc kubenswrapper[4744]: E1201 08:20:18.988711 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.488699156 +0000 UTC m=+151.477757077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.009455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" event={"ID":"b242c027-08d1-4637-9c6a-0b39f07bcf1f","Type":"ContainerStarted","Data":"eebcfce2d57b362a9d464aab31ff4c13f56f4b184c744e82e21767bcd0ec17e6"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.034042 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" event={"ID":"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd","Type":"ContainerStarted","Data":"e461b077f744a0567ef3101a5c054c1dd18b7f00d88200c4d34a21c78226407c"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.035504 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" event={"ID":"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3","Type":"ContainerStarted","Data":"96dfd7f0bbc0a75dd24dfdfa8425003516f868de6e15233b2e7b38e502585889"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.055705 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" event={"ID":"21c2490b-9fde-43ae-ab99-8b5b3758fdd5","Type":"ContainerStarted","Data":"5546e837f8377bcd250bc06d52beaa1a04c3245124508792a03cab7cd7840956"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.055754 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" event={"ID":"21c2490b-9fde-43ae-ab99-8b5b3758fdd5","Type":"ContainerStarted","Data":"5cad4f69b5a5671d95ea152c9ff6bedf2bf33cd2d6750edd4fbf7d989909a376"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.070871 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.073239 4744 generic.go:334] "Generic (PLEG): container finished" podID="33932716-6af1-411f-88c9-a63a12b4077a" containerID="8be368bc1c6cc4e003cb7d1ff2c9273e72caf5fcda37b226feac94774529fc27" exitCode=0 Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.073381 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" event={"ID":"33932716-6af1-411f-88c9-a63a12b4077a","Type":"ContainerDied","Data":"8be368bc1c6cc4e003cb7d1ff2c9273e72caf5fcda37b226feac94774529fc27"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.073519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" event={"ID":"33932716-6af1-411f-88c9-a63a12b4077a","Type":"ContainerStarted","Data":"8e619cece90dbbc1cb3c3abc6b4fb9f77097b7b1d5461f5c862a47575fb827af"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.097894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.099032 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.599014797 +0000 UTC m=+151.588072708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.117954 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" event={"ID":"075bdaa2-f2c7-4c5a-b0ac-09da9c149141","Type":"ContainerStarted","Data":"6b46588751c3b427ef6158fddcdbc032fd00ac1076e298025ff77078e38f4304"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.117995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" event={"ID":"075bdaa2-f2c7-4c5a-b0ac-09da9c149141","Type":"ContainerStarted","Data":"8e84645e8f856f467106d24c48ae245f8c3cb10d9807cbfa550b1d3f1b7e826f"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.123528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" event={"ID":"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61","Type":"ContainerStarted","Data":"2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.124908 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.136842 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.149976 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" event={"ID":"41e0738b-fa8f-425b-8657-452c59c0b8fc","Type":"ContainerStarted","Data":"7d7926fccee4d951acda30677f2b4ecada5863f6d252421ac0389d0ce618833f"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.150024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" event={"ID":"41e0738b-fa8f-425b-8657-452c59c0b8fc","Type":"ContainerStarted","Data":"e8b8ff3e72faef7e2d107b9fb64d859bfe7da24a05fc1e5c06d09e6010584e40"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.165866 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" event={"ID":"9afd374c-6bde-459d-9c6c-61d7ed9cad95","Type":"ContainerStarted","Data":"5a64b97e34ecb2122a4f910dc35a67f74faa78f6f29b2f98ba8ca8755b75faf7"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.201490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.257400 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.757370778 +0000 UTC m=+151.746428699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.287869 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swh9k" podStartSLOduration=126.287796308 podStartE2EDuration="2m6.287796308s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.276264466 +0000 UTC m=+151.265322407" watchObservedRunningTime="2025-12-01 08:20:19.287796308 +0000 UTC m=+151.276854229" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.288348 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-k997m" podStartSLOduration=126.288341683 podStartE2EDuration="2m6.288341683s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.184944946 +0000 UTC m=+151.174002867" watchObservedRunningTime="2025-12-01 08:20:19.288341683 +0000 UTC m=+151.277399604" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.312611 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" event={"ID":"948afa07-87fa-4c8d-a26d-5b9b8d3b156a","Type":"ContainerStarted","Data":"2787b08f1c7d6efadb919e2b8925c3377a2bb9de24214ca1e3cf5962992643f2"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.312708 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" event={"ID":"948afa07-87fa-4c8d-a26d-5b9b8d3b156a","Type":"ContainerStarted","Data":"de8ebdd6b6c3c4b8aa46fafa2efde83ffb9b981e91b0b075aa62ca997016b8b3"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.313045 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.313629 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.316589 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.816562071 +0000 UTC m=+151.805619992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.317588 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qpjtj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.317624 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" podUID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.339710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" event={"ID":"9ad71bbd-aa2a-4637-b894-d0b384822a2f","Type":"ContainerStarted","Data":"4698b293921dc67f080b2a7f027c901250ce302f357e4790e29d17cae28284b9"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.410910 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" event={"ID":"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1","Type":"ContainerStarted","Data":"a131045e0268255a845ee7bcbf5dd44360134b2028bb2c9b0a6783caa0081594"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.417351 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.424748 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:19.924729331 +0000 UTC m=+151.913787252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.463019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" event={"ID":"f71e09c5-239a-4982-ab7d-93c6d647048a","Type":"ContainerStarted","Data":"1d0089b2571506b9421d445488f41e0673c1029e8d2ff23d94ab4dd044d69545"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.463058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" event={"ID":"f71e09c5-239a-4982-ab7d-93c6d647048a","Type":"ContainerStarted","Data":"cf19bc6594afd1f982ece7a7932afa19f02719ef672e7b3ef2934089556123df"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.494391 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" event={"ID":"90b58c6b-55cb-48fd-bdca-1cf449efa230","Type":"ContainerStarted","Data":"6718dbffaab004372d79d62fe49de10f58198627d86575d33e46338ee1f25f7a"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.503139 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" event={"ID":"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b","Type":"ContainerStarted","Data":"b6477076d7fdedee766142dc28fdff006d85adc4e404f901c25853a7bae8ac2d"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.503337 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" event={"ID":"c2a8cae3-3fa3-449c-b339-5fe1eb9afa4b","Type":"ContainerStarted","Data":"0cd2c1d65f138847547f38f2d2f992e53e35b89451e574ce4afb1c1f06c2daa8"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.504359 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.510224 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" podStartSLOduration=126.510200807 podStartE2EDuration="2m6.510200807s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.509772926 +0000 UTC m=+151.498830857" watchObservedRunningTime="2025-12-01 08:20:19.510200807 +0000 UTC m=+151.499258728" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.546275 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.547558 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.04752627 +0000 UTC m=+152.036584191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.552675 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.553210 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.053195168 +0000 UTC m=+152.042253089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.556145 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" event={"ID":"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa","Type":"ContainerStarted","Data":"a7cea3a9d801f2d1a7646865a00ce296e7aba3e6b0035b35853de87e14388c56"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.556185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" event={"ID":"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa","Type":"ContainerStarted","Data":"b0f78b77ed471fc698ebd62e905f7bc196be3a33fb7b3d0b0c8066a3c4145e74"} Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.572622 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.654497 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.656150 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.156128592 +0000 UTC m=+152.145186513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.679160 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xstng" podStartSLOduration=126.679135914 podStartE2EDuration="2m6.679135914s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.583516505 +0000 UTC m=+151.572574416" watchObservedRunningTime="2025-12-01 08:20:19.679135914 +0000 UTC m=+151.668193835" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.680033 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" podStartSLOduration=126.680028429 podStartE2EDuration="2m6.680028429s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.660672699 +0000 UTC m=+151.649730620" watchObservedRunningTime="2025-12-01 08:20:19.680028429 +0000 UTC m=+151.669086340" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.710033 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" podStartSLOduration=126.710012746 podStartE2EDuration="2m6.710012746s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.709033029 +0000 UTC m=+151.698090950" watchObservedRunningTime="2025-12-01 08:20:19.710012746 +0000 UTC m=+151.699070667" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.755156 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9rr8l" podStartSLOduration=126.755129926 podStartE2EDuration="2m6.755129926s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.74417954 +0000 UTC m=+151.733237461" watchObservedRunningTime="2025-12-01 08:20:19.755129926 +0000 UTC m=+151.744187847" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.782975 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.789605 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.289576458 +0000 UTC m=+152.278634379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.792701 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-72sx7" podStartSLOduration=126.792674174 podStartE2EDuration="2m6.792674174s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.7900039 +0000 UTC m=+151.779061821" watchObservedRunningTime="2025-12-01 08:20:19.792674174 +0000 UTC m=+151.781732095" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.854241 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" podStartSLOduration=126.854223923 podStartE2EDuration="2m6.854223923s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.852892276 +0000 UTC m=+151.841950197" watchObservedRunningTime="2025-12-01 08:20:19.854223923 +0000 UTC m=+151.843281834" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.903821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:19 crc kubenswrapper[4744]: E1201 08:20:19.904299 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.404275811 +0000 UTC m=+152.393333732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.908185 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:19 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:19 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:19 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.908709 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:19 crc kubenswrapper[4744]: I1201 08:20:19.956086 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" podStartSLOduration=126.956052296 podStartE2EDuration="2m6.956052296s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:19.937164359 +0000 UTC m=+151.926222280" watchObservedRunningTime="2025-12-01 08:20:19.956052296 +0000 UTC m=+151.945110217" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.011695 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.012180 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.512163193 +0000 UTC m=+152.501221114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.018289 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dmlw7" podStartSLOduration=127.018265673 podStartE2EDuration="2m7.018265673s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:20.014627972 +0000 UTC m=+152.003685893" watchObservedRunningTime="2025-12-01 08:20:20.018265673 +0000 UTC m=+152.007323594" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.114838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.115464 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.615443887 +0000 UTC m=+152.604501808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.217521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.217992 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.717976699 +0000 UTC m=+152.707034620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.305713 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p5bxp" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.320673 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.321535 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.82151057 +0000 UTC m=+152.810568491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.321628 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.322100 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.822079456 +0000 UTC m=+152.811137377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.423387 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.423878 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:20.923855798 +0000 UTC m=+152.912913719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.528910 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.529405 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.029385564 +0000 UTC m=+153.018443485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.625785 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-wvscw" event={"ID":"b2c80d56-3d5a-4ed7-ba9f-a5b80dbd430e","Type":"ContainerStarted","Data":"1fc3787d22a0efdfce3f0a5d97f078d1f4f2772cb2c5d203598f9acafc6fc51a"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.630916 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.632096 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.132079392 +0000 UTC m=+153.121137313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.642813 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5kq46" event={"ID":"66a72013-6e6c-4380-9f5a-c306156e4705","Type":"ContainerStarted","Data":"59fd9a4be9affa89ededed3a63a8b30046347404554def7a2a70e26b30454548"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.659629 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" event={"ID":"c0003968-eec0-455a-ae75-e57629ff675d","Type":"ContainerStarted","Data":"a2c7418daec94f11c7d6a094c929c3f39c09646f0aa97a16702ccc98caf8dedd"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.659694 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" event={"ID":"c0003968-eec0-455a-ae75-e57629ff675d","Type":"ContainerStarted","Data":"6e6baaa3b872b2931bbac8684200e215d507bef7949775459a2900fbf9ac0eac"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.670249 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-wvscw" podStartSLOduration=8.670229667 podStartE2EDuration="8.670229667s" podCreationTimestamp="2025-12-01 08:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:20.666890674 +0000 UTC m=+152.655948595" watchObservedRunningTime="2025-12-01 08:20:20.670229667 +0000 UTC m=+152.659287588" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.688358 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ad71bbd-aa2a-4637-b894-d0b384822a2f" containerID="1d99f253a969970b47dd3129f3f5402d972ff82cd792599661ad4dbb6461c5fe" exitCode=0 Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.688498 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" event={"ID":"9ad71bbd-aa2a-4637-b894-d0b384822a2f","Type":"ContainerDied","Data":"1d99f253a969970b47dd3129f3f5402d972ff82cd792599661ad4dbb6461c5fe"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.716715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" event={"ID":"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1","Type":"ContainerStarted","Data":"0b0826fd0da78cb8f6af09d538d3455fe889459f7328d6a444d57a971f0edc32"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.733131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.745054 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.245030465 +0000 UTC m=+153.234088386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.745855 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" event={"ID":"9a9c9a23-151c-41ce-9a1b-57adbd823c1e","Type":"ContainerStarted","Data":"69ee4ddb962838e01582c2ccf8abc5919cbca2130ae3d4361e7073b2b8abf416"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.766215 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" event={"ID":"33932716-6af1-411f-88c9-a63a12b4077a","Type":"ContainerStarted","Data":"f3835d0cdf6e64bf32147b9601f7aca2224e2daca2b1523afec06b76987bb333"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.804076 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gc7mr" podStartSLOduration=127.804058064 podStartE2EDuration="2m7.804058064s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:20.791956556 +0000 UTC m=+152.781014477" watchObservedRunningTime="2025-12-01 08:20:20.804058064 +0000 UTC m=+152.793115985" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.810913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" event={"ID":"f4bc81ee-8f2e-4ba3-a347-ea31b3e87cdd","Type":"ContainerStarted","Data":"cf1c25a618a4843e5499af5b43618561f162c94a9d7b314ef1af67e0e6260be8"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.811891 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.824236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" event={"ID":"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3","Type":"ContainerStarted","Data":"1c84030850d7c387ba2b783090dbdd76b32f8ba6415f8fb7eb24e007f6a6e33a"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.835136 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.836194 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.33614883 +0000 UTC m=+153.325206911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.841993 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.852534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" event={"ID":"b9b5a1f5-a8ec-4faf-928c-a00ed84cdcfa","Type":"ContainerStarted","Data":"e9c9b7d06b17cfcb28509112dbfc263cbd5ed57d4dde796dde5d289f468b6dac"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.857709 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" podStartSLOduration=127.857685031 podStartE2EDuration="2m7.857685031s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:20.85585018 +0000 UTC m=+152.844908101" watchObservedRunningTime="2025-12-01 08:20:20.857685031 +0000 UTC m=+152.846742952" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.875572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" event={"ID":"1cea4783-96b7-41bd-8a1f-3893008364a2","Type":"ContainerStarted","Data":"c779017f1142ff9506c4db00712c25e46848c1df885c2d0aaf65a5806eca8bee"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.890776 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:20 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:20 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:20 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.890864 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.891346 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" event={"ID":"17390854-a3b3-4906-b28b-5a84a5064964","Type":"ContainerStarted","Data":"c8c4434757ca1161288c04b262942f2565e474a3b7444b06c08b2902c04b6ea2"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.891396 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" event={"ID":"17390854-a3b3-4906-b28b-5a84a5064964","Type":"ContainerStarted","Data":"53e5e40204ea801ba359aa522e60d32d43210f4bc52b0a27758da82d92534776"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.921656 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" event={"ID":"9afd374c-6bde-459d-9c6c-61d7ed9cad95","Type":"ContainerStarted","Data":"20093a3b09ecaf0358e6643b51b4fb247e1c4a00de5faa3a23da4d529e5d1a5f"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.921735 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" event={"ID":"9afd374c-6bde-459d-9c6c-61d7ed9cad95","Type":"ContainerStarted","Data":"b7f9bdb2c16e1d5837dfb0952ee4e6766250a1c9fa7093bda266af9a9e758f1e"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.936850 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:20 crc kubenswrapper[4744]: E1201 08:20:20.937325 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.437305094 +0000 UTC m=+153.426363005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.941333 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" podStartSLOduration=127.941309766 podStartE2EDuration="2m7.941309766s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:20.940148033 +0000 UTC m=+152.929205954" watchObservedRunningTime="2025-12-01 08:20:20.941309766 +0000 UTC m=+152.930367687" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.941982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" event={"ID":"b242c027-08d1-4637-9c6a-0b39f07bcf1f","Type":"ContainerStarted","Data":"776522fb96dca6fb88c5d34395abac20627614c3c8ee445b56e6bd56b904a993"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.942032 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" event={"ID":"b242c027-08d1-4637-9c6a-0b39f07bcf1f","Type":"ContainerStarted","Data":"f02b0a1d8c6968f9403f4ded23c19b56ad94eec7b1a52e6e51587999cb524ce0"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.991782 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ml6wn" event={"ID":"d69f65ac-9803-46d7-a064-4ee35ffa1877","Type":"ContainerStarted","Data":"e6596f948ddd71956ab09a8a76f62b5b9491cc5bc8b07af9042701cdb572d01e"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.991832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ml6wn" event={"ID":"d69f65ac-9803-46d7-a064-4ee35ffa1877","Type":"ContainerStarted","Data":"8b13af44817a9fa408f16673c5c8a12c358fb99dff770c2bffac111147afe299"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.991847 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.998563 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" event={"ID":"1dfa02e3-316f-497a-a7d7-ad6d9315efa6","Type":"ContainerStarted","Data":"16dc99bb076e74553c39c26fabff1e65b27e942ff54bed326ebab851e209fbcd"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.998610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" event={"ID":"1dfa02e3-316f-497a-a7d7-ad6d9315efa6","Type":"ContainerStarted","Data":"9d97e9f029969f75e9f410b9e3cebb383df4f6d0687fe0caf9d396b1f9af8431"} Dec 01 08:20:20 crc kubenswrapper[4744]: I1201 08:20:20.999365 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.014429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nl48d" event={"ID":"90b58c6b-55cb-48fd-bdca-1cf449efa230","Type":"ContainerStarted","Data":"ed8e9b245aa3b5e38f820f68647a0a61294a6794a5d3ae6ef27b854aacce3bd6"} Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.017703 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-mq727 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.017752 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mq727" podUID="5b7ea6bf-6f2c-483e-b37b-9032d56018ee" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.018254 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qpjtj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.018281 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" podUID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.043894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.043976 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.543960172 +0000 UTC m=+153.533018093 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.045391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.052176 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.552162971 +0000 UTC m=+153.541220892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.148196 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.148722 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.648699786 +0000 UTC m=+153.637757707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.250515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.250995 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.750980922 +0000 UTC m=+153.740038843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.296145 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-pnhzx" podStartSLOduration=128.296125323 podStartE2EDuration="2m8.296125323s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.179859496 +0000 UTC m=+153.168917417" watchObservedRunningTime="2025-12-01 08:20:21.296125323 +0000 UTC m=+153.285183244" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.298024 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" podStartSLOduration=128.298018645 podStartE2EDuration="2m8.298018645s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.295637299 +0000 UTC m=+153.284695220" watchObservedRunningTime="2025-12-01 08:20:21.298018645 +0000 UTC m=+153.287076566" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.352063 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.352238 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.852213509 +0000 UTC m=+153.841271430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.352496 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.352930 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.852911378 +0000 UTC m=+153.841969299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.362654 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rhrgq" podStartSLOduration=128.36263572 podStartE2EDuration="2m8.36263572s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.362518706 +0000 UTC m=+153.351576627" watchObservedRunningTime="2025-12-01 08:20:21.36263572 +0000 UTC m=+153.351693641" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.451903 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2ghv7" podStartSLOduration=128.451885341 podStartE2EDuration="2m8.451885341s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.450965635 +0000 UTC m=+153.440023556" watchObservedRunningTime="2025-12-01 08:20:21.451885341 +0000 UTC m=+153.440943262" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.453319 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" podStartSLOduration=128.453314951 podStartE2EDuration="2m8.453314951s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.417210843 +0000 UTC m=+153.406268764" watchObservedRunningTime="2025-12-01 08:20:21.453314951 +0000 UTC m=+153.442372872" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.453787 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.453953 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.953927988 +0000 UTC m=+153.942985909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.454204 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.454636 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:21.954616037 +0000 UTC m=+153.943674138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.492196 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-ml6wn" podStartSLOduration=9.492179546 podStartE2EDuration="9.492179546s" podCreationTimestamp="2025-12-01 08:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.489872321 +0000 UTC m=+153.478930242" watchObservedRunningTime="2025-12-01 08:20:21.492179546 +0000 UTC m=+153.481237467" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.537634 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fh8z6" podStartSLOduration=128.537618654 podStartE2EDuration="2m8.537618654s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.5353234 +0000 UTC m=+153.524381341" watchObservedRunningTime="2025-12-01 08:20:21.537618654 +0000 UTC m=+153.526676575" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.555030 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.555613 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.055589906 +0000 UTC m=+154.044647827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.593274 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.593881 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.597676 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.597924 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.632955 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8jr2s" podStartSLOduration=128.632926696 podStartE2EDuration="2m8.632926696s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.597900338 +0000 UTC m=+153.586958259" watchObservedRunningTime="2025-12-01 08:20:21.632926696 +0000 UTC m=+153.621984617" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.633199 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7th4n" podStartSLOduration=128.633194163 podStartE2EDuration="2m8.633194163s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:21.631007712 +0000 UTC m=+153.620065633" watchObservedRunningTime="2025-12-01 08:20:21.633194163 +0000 UTC m=+153.622252084" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.640979 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.657224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.657299 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.657338 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.657654 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.157637145 +0000 UTC m=+154.146695066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.758354 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.758710 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.258671367 +0000 UTC m=+154.247729458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.758841 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.758907 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.758982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.781134 4744 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.800461 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.860064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.860517 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.36050509 +0000 UTC m=+154.349563011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.900743 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:21 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:21 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:21 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.900873 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.913054 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.930258 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pnzr5"] Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.945832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.951792 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.961821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.962008 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-utilities\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.962092 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdklm\" (UniqueName: \"kubernetes.io/projected/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-kube-api-access-rdklm\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.962117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-catalog-content\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:21 crc kubenswrapper[4744]: E1201 08:20:21.962297 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.462270161 +0000 UTC m=+154.451328082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:21 crc kubenswrapper[4744]: I1201 08:20:21.967709 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pnzr5"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.064018 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" event={"ID":"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1","Type":"ContainerStarted","Data":"807fbb2d5cd5df7c9927389b8ba709450a7197f3c9c8ddfa30b76c1e7fcec6c8"} Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.066055 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.066095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdklm\" (UniqueName: \"kubernetes.io/projected/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-kube-api-access-rdklm\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.066120 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-catalog-content\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.066171 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-utilities\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.066574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-utilities\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:22 crc kubenswrapper[4744]: E1201 08:20:22.066775 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.566760929 +0000 UTC m=+154.555818850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.067048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-catalog-content\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.082903 4744 generic.go:334] "Generic (PLEG): container finished" podID="9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" containerID="1c84030850d7c387ba2b783090dbdd76b32f8ba6415f8fb7eb24e007f6a6e33a" exitCode=0 Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.082986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" event={"ID":"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3","Type":"ContainerDied","Data":"1c84030850d7c387ba2b783090dbdd76b32f8ba6415f8fb7eb24e007f6a6e33a"} Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.100546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdklm\" (UniqueName: \"kubernetes.io/projected/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-kube-api-access-rdklm\") pod \"certified-operators-pnzr5\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.121498 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" event={"ID":"9ad71bbd-aa2a-4637-b894-d0b384822a2f","Type":"ContainerStarted","Data":"4aef56beafe366728dc9dd949eb0adb39ab4c91ff2b299f2bf88d2dc642c11c5"} Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.126958 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.134529 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pvtt2"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.138634 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.144527 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.158211 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pvtt2"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.169777 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.169875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-catalog-content\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: E1201 08:20:22.170093 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.670040552 +0000 UTC m=+154.659098483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.174322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-utilities\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.178276 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" podStartSLOduration=129.178249332 podStartE2EDuration="2m9.178249332s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:22.177445219 +0000 UTC m=+154.166503130" watchObservedRunningTime="2025-12-01 08:20:22.178249332 +0000 UTC m=+154.167307253" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.194117 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.194333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msgnj\" (UniqueName: \"kubernetes.io/projected/d2e302f9-45f6-4e8e-9c08-984faf8732d5-kube-api-access-msgnj\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: E1201 08:20:22.195239 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.695224256 +0000 UTC m=+154.684282177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.295523 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.296106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msgnj\" (UniqueName: \"kubernetes.io/projected/d2e302f9-45f6-4e8e-9c08-984faf8732d5-kube-api-access-msgnj\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.296134 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-catalog-content\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.296177 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-utilities\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.296585 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-utilities\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: E1201 08:20:22.296655 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.796638127 +0000 UTC m=+154.785696048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.297098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-catalog-content\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.326337 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-md89c"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.327203 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.334090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msgnj\" (UniqueName: \"kubernetes.io/projected/d2e302f9-45f6-4e8e-9c08-984faf8732d5-kube-api-access-msgnj\") pod \"community-operators-pvtt2\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.334260 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.352880 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-md89c"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.399572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-utilities\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.400025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-catalog-content\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.400135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqsdl\" (UniqueName: \"kubernetes.io/projected/e9d0b50b-d228-4113-b88f-3ff873f6367e-kube-api-access-mqsdl\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.400294 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:22 crc kubenswrapper[4744]: E1201 08:20:22.400826 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:20:22.900809526 +0000 UTC m=+154.889867447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l6djk" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.478862 4744 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T08:20:21.781180915Z","Handler":null,"Name":""} Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.518939 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.519156 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-catalog-content\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.519202 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqsdl\" (UniqueName: \"kubernetes.io/projected/e9d0b50b-d228-4113-b88f-3ff873f6367e-kube-api-access-mqsdl\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.519267 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-utilities\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.519673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-utilities\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: E1201 08:20:22.519739 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:20:23.019725356 +0000 UTC m=+155.008783277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.519943 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-catalog-content\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.536735 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.537053 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v5s7x"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.537201 4744 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.537243 4744 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.538536 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.561362 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqsdl\" (UniqueName: \"kubernetes.io/projected/e9d0b50b-d228-4113-b88f-3ff873f6367e-kube-api-access-mqsdl\") pod \"certified-operators-md89c\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.575222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5s7x"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.636808 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.673289 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.677891 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.744018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-catalog-content\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.744100 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-utilities\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.744322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cghb\" (UniqueName: \"kubernetes.io/projected/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-kube-api-access-5cghb\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.749820 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.749872 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.848143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cghb\" (UniqueName: \"kubernetes.io/projected/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-kube-api-access-5cghb\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.848244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-catalog-content\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.848300 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-utilities\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.848883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-utilities\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.848985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-catalog-content\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.880592 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cghb\" (UniqueName: \"kubernetes.io/projected/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-kube-api-access-5cghb\") pod \"community-operators-v5s7x\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.894657 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:22 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:22 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:22 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.895162 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.913719 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pnzr5"] Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.931932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:20:22 crc kubenswrapper[4744]: I1201 08:20:22.964223 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l6djk\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.005196 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pvtt2"] Dec 01 08:20:23 crc kubenswrapper[4744]: W1201 08:20:23.013300 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2e302f9_45f6_4e8e_9c08_984faf8732d5.slice/crio-04bcc6af421a415dcfb1e7dc3df554233069eeab318ffccad577d330678eac99 WatchSource:0}: Error finding container 04bcc6af421a415dcfb1e7dc3df554233069eeab318ffccad577d330678eac99: Status 404 returned error can't find the container with id 04bcc6af421a415dcfb1e7dc3df554233069eeab318ffccad577d330678eac99 Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.061655 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.095657 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.100293 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-md89c"] Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.133067 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnzr5" event={"ID":"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b","Type":"ContainerStarted","Data":"5045d5b2f6159e439f505fa8b1bd8c65cbd671a3807f58e103c3afcd64f6d0fe"} Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.138403 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" event={"ID":"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1","Type":"ContainerStarted","Data":"5dc8504303873b411be80c995e195bf7354000bc0d862c95788d04ebc92624dd"} Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.141374 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"131d5b2c-08e1-4bd7-98ca-24d329f4a78d","Type":"ContainerStarted","Data":"1f7461d8cb1e77f300a30ca87185ff8c5e82748edfad6f81282612e23297eea4"} Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.142883 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvtt2" event={"ID":"d2e302f9-45f6-4e8e-9c08-984faf8732d5","Type":"ContainerStarted","Data":"04bcc6af421a415dcfb1e7dc3df554233069eeab318ffccad577d330678eac99"} Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.144113 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md89c" event={"ID":"e9d0b50b-d228-4113-b88f-3ff873f6367e","Type":"ContainerStarted","Data":"8b77e42c89881ff17a162f8ba771f893e1bf5917e970c08223dca23056f093a3"} Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.148849 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-44bgh" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.208330 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5s7x"] Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.224356 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:23 crc kubenswrapper[4744]: W1201 08:20:23.312515 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c6dc85b_81b3_42a1_92ea_93977c98dc1b.slice/crio-7c22a82ccfd773267c170ecf9f30823bea00be206f45011a21893a5c5bc89852 WatchSource:0}: Error finding container 7c22a82ccfd773267c170ecf9f30823bea00be206f45011a21893a5c5bc89852: Status 404 returned error can't find the container with id 7c22a82ccfd773267c170ecf9f30823bea00be206f45011a21893a5c5bc89852 Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.487946 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.574798 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l6djk"] Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.675127 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ksch\" (UniqueName: \"kubernetes.io/projected/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-kube-api-access-8ksch\") pod \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.675289 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-secret-volume\") pod \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.675347 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-config-volume\") pod \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\" (UID: \"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3\") " Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.676717 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-config-volume" (OuterVolumeSpecName: "config-volume") pod "9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" (UID: "9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.683627 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-kube-api-access-8ksch" (OuterVolumeSpecName: "kube-api-access-8ksch") pod "9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" (UID: "9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3"). InnerVolumeSpecName "kube-api-access-8ksch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.687122 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" (UID: "9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:20:23 crc kubenswrapper[4744]: W1201 08:20:23.696215 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5516996f_81ea_44d0_b5a1_5af61f00af8c.slice/crio-d90e8b978ecf142bc1d3be88896d315f65bc8d44faf70adb7f33710cceeab76c WatchSource:0}: Error finding container d90e8b978ecf142bc1d3be88896d315f65bc8d44faf70adb7f33710cceeab76c: Status 404 returned error can't find the container with id d90e8b978ecf142bc1d3be88896d315f65bc8d44faf70adb7f33710cceeab76c Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.776728 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.776768 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.776782 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ksch\" (UniqueName: \"kubernetes.io/projected/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3-kube-api-access-8ksch\") on node \"crc\" DevicePath \"\"" Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.881075 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:23 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:23 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:23 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:23 crc kubenswrapper[4744]: I1201 08:20:23.881151 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.115345 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6rscr"] Dec 01 08:20:24 crc kubenswrapper[4744]: E1201 08:20:24.115918 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" containerName="collect-profiles" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.115932 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" containerName="collect-profiles" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.116041 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" containerName="collect-profiles" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.116784 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.119535 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.136258 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rscr"] Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.180608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxd8c\" (UniqueName: \"kubernetes.io/projected/862392cd-33ed-4e07-8ea5-eef33e192765-kube-api-access-dxd8c\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.180672 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-catalog-content\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.180760 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-utilities\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.181517 4744 generic.go:334] "Generic (PLEG): container finished" podID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerID="63bb2f481bdf53d281e5dea3a0480333e24c406e8a2ae53ee2f2b2cc08dc00c2" exitCode=0 Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.181639 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnzr5" event={"ID":"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b","Type":"ContainerDied","Data":"63bb2f481bdf53d281e5dea3a0480333e24c406e8a2ae53ee2f2b2cc08dc00c2"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.183177 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" event={"ID":"5516996f-81ea-44d0-b5a1-5af61f00af8c","Type":"ContainerStarted","Data":"3ba489857f29415fc09c5af9fb635237717ee48896e20f7fca1b3e5e69eca709"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.183201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" event={"ID":"5516996f-81ea-44d0-b5a1-5af61f00af8c","Type":"ContainerStarted","Data":"d90e8b978ecf142bc1d3be88896d315f65bc8d44faf70adb7f33710cceeab76c"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.183834 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.184161 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.189886 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" event={"ID":"7aaa5201-50c4-4ceb-90b9-f6ddbfebb0d1","Type":"ContainerStarted","Data":"bc406c21f588344d45100de4037564deaca4ce1c01349ce3e384e98a1625837a"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.192086 4744 generic.go:334] "Generic (PLEG): container finished" podID="131d5b2c-08e1-4bd7-98ca-24d329f4a78d" containerID="13db89c910d3d25c5f62aa9445a8a38144ee75c7b9b9f599ba7ed56a3853818b" exitCode=0 Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.192125 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"131d5b2c-08e1-4bd7-98ca-24d329f4a78d","Type":"ContainerDied","Data":"13db89c910d3d25c5f62aa9445a8a38144ee75c7b9b9f599ba7ed56a3853818b"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.193983 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.194005 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55" event={"ID":"9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3","Type":"ContainerDied","Data":"96dfd7f0bbc0a75dd24dfdfa8425003516f868de6e15233b2e7b38e502585889"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.194035 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96dfd7f0bbc0a75dd24dfdfa8425003516f868de6e15233b2e7b38e502585889" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.196087 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerID="f8955fd17e87e863e87f3143141d3542f40c2f19348f5913c14e3ddd3a6ba04f" exitCode=0 Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.196170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvtt2" event={"ID":"d2e302f9-45f6-4e8e-9c08-984faf8732d5","Type":"ContainerDied","Data":"f8955fd17e87e863e87f3143141d3542f40c2f19348f5913c14e3ddd3a6ba04f"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.212385 4744 generic.go:334] "Generic (PLEG): container finished" podID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerID="04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b" exitCode=0 Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.212469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md89c" event={"ID":"e9d0b50b-d228-4113-b88f-3ff873f6367e","Type":"ContainerDied","Data":"04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.216692 4744 generic.go:334] "Generic (PLEG): container finished" podID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerID="a1eb48880392d3e966454f7be898e6b98f8c4bcdea3113dc2848febc78cb18e7" exitCode=0 Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.216784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5s7x" event={"ID":"2c6dc85b-81b3-42a1-92ea-93977c98dc1b","Type":"ContainerDied","Data":"a1eb48880392d3e966454f7be898e6b98f8c4bcdea3113dc2848febc78cb18e7"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.216851 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5s7x" event={"ID":"2c6dc85b-81b3-42a1-92ea-93977c98dc1b","Type":"ContainerStarted","Data":"7c22a82ccfd773267c170ecf9f30823bea00be206f45011a21893a5c5bc89852"} Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.257688 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" podStartSLOduration=131.257665992 podStartE2EDuration="2m11.257665992s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:24.253369512 +0000 UTC m=+156.242427443" watchObservedRunningTime="2025-12-01 08:20:24.257665992 +0000 UTC m=+156.246723913" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.282000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxd8c\" (UniqueName: \"kubernetes.io/projected/862392cd-33ed-4e07-8ea5-eef33e192765-kube-api-access-dxd8c\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.282272 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-catalog-content\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.282852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-catalog-content\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.282942 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-utilities\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.284172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-utilities\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.313992 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxd8c\" (UniqueName: \"kubernetes.io/projected/862392cd-33ed-4e07-8ea5-eef33e192765-kube-api-access-dxd8c\") pod \"redhat-marketplace-6rscr\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.318225 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.360727 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wr5kd" podStartSLOduration=12.360706159 podStartE2EDuration="12.360706159s" podCreationTimestamp="2025-12-01 08:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:24.357229361 +0000 UTC m=+156.346287292" watchObservedRunningTime="2025-12-01 08:20:24.360706159 +0000 UTC m=+156.349764080" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.434101 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.510065 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ngq48"] Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.514013 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.525225 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngq48"] Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.694091 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rscr"] Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.703041 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-catalog-content\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.703096 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-utilities\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.703160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k7k9\" (UniqueName: \"kubernetes.io/projected/bd79b9ad-243e-470a-8c89-2a05be15e2ab-kube-api-access-9k7k9\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.708867 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.708899 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.714208 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:24 crc kubenswrapper[4744]: W1201 08:20:24.718292 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod862392cd_33ed_4e07_8ea5_eef33e192765.slice/crio-e24adbee2855d9ca517a1ce836111b048190cdb9366826927d2fa54e5285bf54 WatchSource:0}: Error finding container e24adbee2855d9ca517a1ce836111b048190cdb9366826927d2fa54e5285bf54: Status 404 returned error can't find the container with id e24adbee2855d9ca517a1ce836111b048190cdb9366826927d2fa54e5285bf54 Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.804445 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-catalog-content\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.804513 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-utilities\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.804596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k7k9\" (UniqueName: \"kubernetes.io/projected/bd79b9ad-243e-470a-8c89-2a05be15e2ab-kube-api-access-9k7k9\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.805979 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-catalog-content\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.806008 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-utilities\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.828058 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k7k9\" (UniqueName: \"kubernetes.io/projected/bd79b9ad-243e-470a-8c89-2a05be15e2ab-kube-api-access-9k7k9\") pod \"redhat-marketplace-ngq48\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.839376 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.840156 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.840506 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.841874 4744 patch_prober.go:28] interesting pod/console-f9d7485db-nmgd9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.841925 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nmgd9" podUID="3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.879360 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:24 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:24 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:24 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:24 crc kubenswrapper[4744]: I1201 08:20:24.879426 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.115664 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gs9pb"] Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.116856 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.118897 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.125446 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gs9pb"] Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.178466 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-mq727 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.178528 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mq727" podUID="5b7ea6bf-6f2c-483e-b37b-9032d56018ee" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.178479 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-mq727 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.178612 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mq727" podUID="5b7ea6bf-6f2c-483e-b37b-9032d56018ee" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.210728 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-utilities\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.210788 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5xgf\" (UniqueName: \"kubernetes.io/projected/fe4049b3-153a-45d1-8847-a51f91053f77-kube-api-access-z5xgf\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.210821 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-catalog-content\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.224067 4744 generic.go:334] "Generic (PLEG): container finished" podID="862392cd-33ed-4e07-8ea5-eef33e192765" containerID="7155be193e36537589e6d80d1e8fb08b398ef3f23e04d6b349147069af172b83" exitCode=0 Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.225326 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rscr" event={"ID":"862392cd-33ed-4e07-8ea5-eef33e192765","Type":"ContainerDied","Data":"7155be193e36537589e6d80d1e8fb08b398ef3f23e04d6b349147069af172b83"} Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.225364 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rscr" event={"ID":"862392cd-33ed-4e07-8ea5-eef33e192765","Type":"ContainerStarted","Data":"e24adbee2855d9ca517a1ce836111b048190cdb9366826927d2fa54e5285bf54"} Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.229852 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-wl2vv" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.313639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-utilities\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.313718 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5xgf\" (UniqueName: \"kubernetes.io/projected/fe4049b3-153a-45d1-8847-a51f91053f77-kube-api-access-z5xgf\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.313750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-catalog-content\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.314465 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-catalog-content\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.314673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-utilities\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.327708 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngq48"] Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.372107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5xgf\" (UniqueName: \"kubernetes.io/projected/fe4049b3-153a-45d1-8847-a51f91053f77-kube-api-access-z5xgf\") pod \"redhat-operators-gs9pb\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.447001 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.514566 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-spj2t"] Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.515696 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.531074 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-spj2t"] Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.576069 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.604677 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.626009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-catalog-content\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.626258 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2vxb\" (UniqueName: \"kubernetes.io/projected/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-kube-api-access-h2vxb\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.626288 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-utilities\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.657253 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.657308 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.667579 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.727272 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kubelet-dir\") pod \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.727322 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kube-api-access\") pod \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\" (UID: \"131d5b2c-08e1-4bd7-98ca-24d329f4a78d\") " Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.727467 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-catalog-content\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.727487 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2vxb\" (UniqueName: \"kubernetes.io/projected/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-kube-api-access-h2vxb\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.727521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-utilities\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.728251 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "131d5b2c-08e1-4bd7-98ca-24d329f4a78d" (UID: "131d5b2c-08e1-4bd7-98ca-24d329f4a78d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.729529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-catalog-content\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.729760 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-utilities\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.741848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "131d5b2c-08e1-4bd7-98ca-24d329f4a78d" (UID: "131d5b2c-08e1-4bd7-98ca-24d329f4a78d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.775157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2vxb\" (UniqueName: \"kubernetes.io/projected/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-kube-api-access-h2vxb\") pod \"redhat-operators-spj2t\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.828817 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.828854 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/131d5b2c-08e1-4bd7-98ca-24d329f4a78d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.854257 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.876634 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.879637 4744 patch_prober.go:28] interesting pod/router-default-5444994796-q6bv4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:20:25 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Dec 01 08:20:25 crc kubenswrapper[4744]: [+]process-running ok Dec 01 08:20:25 crc kubenswrapper[4744]: healthz check failed Dec 01 08:20:25 crc kubenswrapper[4744]: I1201 08:20:25.879714 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q6bv4" podUID="85f3bf48-28bd-4afe-b0d9-a0fbd11de675" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:25.991492 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gs9pb"] Dec 01 08:20:26 crc kubenswrapper[4744]: W1201 08:20:25.993638 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe4049b3_153a_45d1_8847_a51f91053f77.slice/crio-bb41122cb3450bdf6cafcf9935f7417deb02d176125e057767480ea83d05d6ea WatchSource:0}: Error finding container bb41122cb3450bdf6cafcf9935f7417deb02d176125e057767480ea83d05d6ea: Status 404 returned error can't find the container with id bb41122cb3450bdf6cafcf9935f7417deb02d176125e057767480ea83d05d6ea Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.084124 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-spj2t"] Dec 01 08:20:26 crc kubenswrapper[4744]: W1201 08:20:26.118511 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod945f44ee_e2a3_4a1a_a6c0_4cd6b107c482.slice/crio-14d7586e56ff6c3f96bc30c65e6e859101f631914850677ef292d6b1383d800d WatchSource:0}: Error finding container 14d7586e56ff6c3f96bc30c65e6e859101f631914850677ef292d6b1383d800d: Status 404 returned error can't find the container with id 14d7586e56ff6c3f96bc30c65e6e859101f631914850677ef292d6b1383d800d Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.233793 4744 generic.go:334] "Generic (PLEG): container finished" podID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerID="8713cb16920f45df71c3eca2e7514466eaf261e6406a91f8b70cf8affb6421dd" exitCode=0 Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.233869 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngq48" event={"ID":"bd79b9ad-243e-470a-8c89-2a05be15e2ab","Type":"ContainerDied","Data":"8713cb16920f45df71c3eca2e7514466eaf261e6406a91f8b70cf8affb6421dd"} Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.233896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngq48" event={"ID":"bd79b9ad-243e-470a-8c89-2a05be15e2ab","Type":"ContainerStarted","Data":"0c69f76fdb27a92c1d338bbbc19473e41b7bdd8377e188b747719e41880556e1"} Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.245607 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spj2t" event={"ID":"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482","Type":"ContainerStarted","Data":"14d7586e56ff6c3f96bc30c65e6e859101f631914850677ef292d6b1383d800d"} Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.271221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"131d5b2c-08e1-4bd7-98ca-24d329f4a78d","Type":"ContainerDied","Data":"1f7461d8cb1e77f300a30ca87185ff8c5e82748edfad6f81282612e23297eea4"} Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.271260 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f7461d8cb1e77f300a30ca87185ff8c5e82748edfad6f81282612e23297eea4" Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.271313 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.285061 4744 generic.go:334] "Generic (PLEG): container finished" podID="fe4049b3-153a-45d1-8847-a51f91053f77" containerID="b591cfe080e5743432c243050dbb66b23888d0ec6a9fe0c87ae05755d1b1e0da" exitCode=0 Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.304166 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7m2jp" Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.304199 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs9pb" event={"ID":"fe4049b3-153a-45d1-8847-a51f91053f77","Type":"ContainerDied","Data":"b591cfe080e5743432c243050dbb66b23888d0ec6a9fe0c87ae05755d1b1e0da"} Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.304218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs9pb" event={"ID":"fe4049b3-153a-45d1-8847-a51f91053f77","Type":"ContainerStarted","Data":"bb41122cb3450bdf6cafcf9935f7417deb02d176125e057767480ea83d05d6ea"} Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.880876 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:26 crc kubenswrapper[4744]: I1201 08:20:26.883585 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-q6bv4" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.147975 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:20:27 crc kubenswrapper[4744]: E1201 08:20:27.148530 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131d5b2c-08e1-4bd7-98ca-24d329f4a78d" containerName="pruner" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.148603 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="131d5b2c-08e1-4bd7-98ca-24d329f4a78d" containerName="pruner" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.148746 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="131d5b2c-08e1-4bd7-98ca-24d329f4a78d" containerName="pruner" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.149243 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.153666 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.155539 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.160132 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.289699 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec8a8f60-63da-4d68-909c-8a958c4e116d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.289781 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec8a8f60-63da-4d68-909c-8a958c4e116d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.296274 4744 generic.go:334] "Generic (PLEG): container finished" podID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerID="5937f44f65ae6cbd2308697dcd6cb23c925c6f62f7eed4b8a00bbe20231d00f1" exitCode=0 Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.296550 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spj2t" event={"ID":"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482","Type":"ContainerDied","Data":"5937f44f65ae6cbd2308697dcd6cb23c925c6f62f7eed4b8a00bbe20231d00f1"} Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.391356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec8a8f60-63da-4d68-909c-8a958c4e116d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.391441 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec8a8f60-63da-4d68-909c-8a958c4e116d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.391509 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec8a8f60-63da-4d68-909c-8a958c4e116d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.410632 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec8a8f60-63da-4d68-909c-8a958c4e116d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.466540 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:27 crc kubenswrapper[4744]: I1201 08:20:27.841021 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:20:28 crc kubenswrapper[4744]: I1201 08:20:28.312709 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec8a8f60-63da-4d68-909c-8a958c4e116d","Type":"ContainerStarted","Data":"e6d6e01c7ce1ce1da76a076096ca72752d3e1c2cf6c354e084cc232767a294de"} Dec 01 08:20:29 crc kubenswrapper[4744]: I1201 08:20:29.328595 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec8a8f60-63da-4d68-909c-8a958c4e116d","Type":"ContainerStarted","Data":"f135dfb9725b90a69fd2fa59a019ecdd51f2830b938eb52ee1689a31bc108296"} Dec 01 08:20:29 crc kubenswrapper[4744]: I1201 08:20:29.350582 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.350537499 podStartE2EDuration="2.350537499s" podCreationTimestamp="2025-12-01 08:20:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:29.341951509 +0000 UTC m=+161.331009430" watchObservedRunningTime="2025-12-01 08:20:29.350537499 +0000 UTC m=+161.339595440" Dec 01 08:20:30 crc kubenswrapper[4744]: I1201 08:20:30.341905 4744 generic.go:334] "Generic (PLEG): container finished" podID="ec8a8f60-63da-4d68-909c-8a958c4e116d" containerID="f135dfb9725b90a69fd2fa59a019ecdd51f2830b938eb52ee1689a31bc108296" exitCode=0 Dec 01 08:20:30 crc kubenswrapper[4744]: I1201 08:20:30.344594 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec8a8f60-63da-4d68-909c-8a958c4e116d","Type":"ContainerDied","Data":"f135dfb9725b90a69fd2fa59a019ecdd51f2830b938eb52ee1689a31bc108296"} Dec 01 08:20:31 crc kubenswrapper[4744]: I1201 08:20:31.048759 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-ml6wn" Dec 01 08:20:34 crc kubenswrapper[4744]: I1201 08:20:34.879612 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:34 crc kubenswrapper[4744]: I1201 08:20:34.884140 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:20:35 crc kubenswrapper[4744]: I1201 08:20:35.200035 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mq727" Dec 01 08:20:36 crc kubenswrapper[4744]: I1201 08:20:36.249087 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:20:36 crc kubenswrapper[4744]: I1201 08:20:36.256576 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8ac81c1-4bf7-472c-8110-21364dde7fd6-metrics-certs\") pod \"network-metrics-daemon-p489k\" (UID: \"c8ac81c1-4bf7-472c-8110-21364dde7fd6\") " pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:20:36 crc kubenswrapper[4744]: I1201 08:20:36.505256 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p489k" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.089196 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.185850 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec8a8f60-63da-4d68-909c-8a958c4e116d-kubelet-dir\") pod \"ec8a8f60-63da-4d68-909c-8a958c4e116d\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.185933 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec8a8f60-63da-4d68-909c-8a958c4e116d-kube-api-access\") pod \"ec8a8f60-63da-4d68-909c-8a958c4e116d\" (UID: \"ec8a8f60-63da-4d68-909c-8a958c4e116d\") " Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.185958 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec8a8f60-63da-4d68-909c-8a958c4e116d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ec8a8f60-63da-4d68-909c-8a958c4e116d" (UID: "ec8a8f60-63da-4d68-909c-8a958c4e116d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.186345 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec8a8f60-63da-4d68-909c-8a958c4e116d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.191295 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8a8f60-63da-4d68-909c-8a958c4e116d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ec8a8f60-63da-4d68-909c-8a958c4e116d" (UID: "ec8a8f60-63da-4d68-909c-8a958c4e116d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.259969 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p489k"] Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.286949 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec8a8f60-63da-4d68-909c-8a958c4e116d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.395263 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec8a8f60-63da-4d68-909c-8a958c4e116d","Type":"ContainerDied","Data":"e6d6e01c7ce1ce1da76a076096ca72752d3e1c2cf6c354e084cc232767a294de"} Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.395310 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6d6e01c7ce1ce1da76a076096ca72752d3e1c2cf6c354e084cc232767a294de" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.395365 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:20:38 crc kubenswrapper[4744]: I1201 08:20:38.397462 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p489k" event={"ID":"c8ac81c1-4bf7-472c-8110-21364dde7fd6","Type":"ContainerStarted","Data":"17c6b79c93be49995139493e429b82152b6d2ae16892260b98424c6e18e68522"} Dec 01 08:20:39 crc kubenswrapper[4744]: I1201 08:20:39.404707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p489k" event={"ID":"c8ac81c1-4bf7-472c-8110-21364dde7fd6","Type":"ContainerStarted","Data":"345f4035f5fbae6a930041065b8782f56a38774d8460c66ffe1bbcb692934f3f"} Dec 01 08:20:40 crc kubenswrapper[4744]: I1201 08:20:40.411610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p489k" event={"ID":"c8ac81c1-4bf7-472c-8110-21364dde7fd6","Type":"ContainerStarted","Data":"85b99b6df5943933630dde7b4cce5c27f398d2d474948d8cc0f5c2eb6cff8871"} Dec 01 08:20:40 crc kubenswrapper[4744]: I1201 08:20:40.427841 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-p489k" podStartSLOduration=147.427824648 podStartE2EDuration="2m27.427824648s" podCreationTimestamp="2025-12-01 08:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:20:40.425687769 +0000 UTC m=+172.414745700" watchObservedRunningTime="2025-12-01 08:20:40.427824648 +0000 UTC m=+172.416882569" Dec 01 08:20:43 crc kubenswrapper[4744]: I1201 08:20:43.234186 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:20:48 crc kubenswrapper[4744]: I1201 08:20:48.635394 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:20:48 crc kubenswrapper[4744]: I1201 08:20:48.636282 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:20:53 crc kubenswrapper[4744]: E1201 08:20:53.668992 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 08:20:53 crc kubenswrapper[4744]: E1201 08:20:53.669467 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdklm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-pnzr5_openshift-marketplace(6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:53 crc kubenswrapper[4744]: E1201 08:20:53.670796 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-pnzr5" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" Dec 01 08:20:54 crc kubenswrapper[4744]: I1201 08:20:54.593167 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:20:54 crc kubenswrapper[4744]: E1201 08:20:54.607754 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-pnzr5" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" Dec 01 08:20:54 crc kubenswrapper[4744]: E1201 08:20:54.692210 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 08:20:54 crc kubenswrapper[4744]: E1201 08:20:54.692443 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5cghb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-v5s7x_openshift-marketplace(2c6dc85b-81b3-42a1-92ea-93977c98dc1b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:54 crc kubenswrapper[4744]: E1201 08:20:54.694196 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-v5s7x" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" Dec 01 08:20:54 crc kubenswrapper[4744]: E1201 08:20:54.715284 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 08:20:54 crc kubenswrapper[4744]: E1201 08:20:54.715465 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mqsdl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-md89c_openshift-marketplace(e9d0b50b-d228-4113-b88f-3ff873f6367e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:54 crc kubenswrapper[4744]: E1201 08:20:54.716692 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-md89c" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" Dec 01 08:20:55 crc kubenswrapper[4744]: E1201 08:20:55.691818 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-v5s7x" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" Dec 01 08:20:55 crc kubenswrapper[4744]: E1201 08:20:55.692377 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-md89c" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" Dec 01 08:20:55 crc kubenswrapper[4744]: E1201 08:20:55.746546 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 08:20:55 crc kubenswrapper[4744]: E1201 08:20:55.746763 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dxd8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6rscr_openshift-marketplace(862392cd-33ed-4e07-8ea5-eef33e192765): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:55 crc kubenswrapper[4744]: E1201 08:20:55.747924 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6rscr" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" Dec 01 08:20:55 crc kubenswrapper[4744]: I1201 08:20:55.961284 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gjtqv" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.770023 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6rscr" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.855170 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.855300 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msgnj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pvtt2_openshift-marketplace(d2e302f9-45f6-4e8e-9c08-984faf8732d5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.856484 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pvtt2" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.863533 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.863675 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9k7k9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ngq48_openshift-marketplace(bd79b9ad-243e-470a-8c89-2a05be15e2ab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.864775 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ngq48" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.869870 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.869957 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h2vxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-spj2t_openshift-marketplace(945f44ee-e2a3-4a1a-a6c0-4cd6b107c482): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.871103 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-spj2t" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.884472 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.884648 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5xgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gs9pb_openshift-marketplace(fe4049b3-153a-45d1-8847-a51f91053f77): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:20:58 crc kubenswrapper[4744]: E1201 08:20:58.885798 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gs9pb" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" Dec 01 08:20:59 crc kubenswrapper[4744]: E1201 08:20:59.518503 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-pvtt2" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" Dec 01 08:20:59 crc kubenswrapper[4744]: E1201 08:20:59.518574 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ngq48" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" Dec 01 08:20:59 crc kubenswrapper[4744]: E1201 08:20:59.518591 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-spj2t" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" Dec 01 08:20:59 crc kubenswrapper[4744]: E1201 08:20:59.518851 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gs9pb" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.202222 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:21:05 crc kubenswrapper[4744]: E1201 08:21:05.205722 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8a8f60-63da-4d68-909c-8a958c4e116d" containerName="pruner" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.205739 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8a8f60-63da-4d68-909c-8a958c4e116d" containerName="pruner" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.205838 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8a8f60-63da-4d68-909c-8a958c4e116d" containerName="pruner" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.206247 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.213279 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.213514 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.213537 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.353653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.353753 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.454360 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.454466 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.454637 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.479505 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.526569 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:05 crc kubenswrapper[4744]: I1201 08:21:05.848574 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:21:06 crc kubenswrapper[4744]: I1201 08:21:06.547267 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"660aa0b3-48c4-4bfa-b712-f29e62fba0f4","Type":"ContainerStarted","Data":"3c64381cc18e32d0d93506211432b06da2e1ac16cf8af3a9c56d6095ac993193"} Dec 01 08:21:07 crc kubenswrapper[4744]: I1201 08:21:07.553104 4744 generic.go:334] "Generic (PLEG): container finished" podID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerID="16ecd5a22c1907203d773d276bf2b1da861abea3bb645d063710448d6264b066" exitCode=0 Dec 01 08:21:07 crc kubenswrapper[4744]: I1201 08:21:07.553954 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5s7x" event={"ID":"2c6dc85b-81b3-42a1-92ea-93977c98dc1b","Type":"ContainerDied","Data":"16ecd5a22c1907203d773d276bf2b1da861abea3bb645d063710448d6264b066"} Dec 01 08:21:07 crc kubenswrapper[4744]: I1201 08:21:07.558619 4744 generic.go:334] "Generic (PLEG): container finished" podID="660aa0b3-48c4-4bfa-b712-f29e62fba0f4" containerID="a07cd268a052c61b481261b2226903c2c235c5e5cceed632b76cbfdd36607bba" exitCode=0 Dec 01 08:21:07 crc kubenswrapper[4744]: I1201 08:21:07.558646 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"660aa0b3-48c4-4bfa-b712-f29e62fba0f4","Type":"ContainerDied","Data":"a07cd268a052c61b481261b2226903c2c235c5e5cceed632b76cbfdd36607bba"} Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.565787 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5s7x" event={"ID":"2c6dc85b-81b3-42a1-92ea-93977c98dc1b","Type":"ContainerStarted","Data":"0a708ed95c06d19828bbe309d1645da3d9a2ad36bd51a83d6463fe42f52af9bc"} Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.589482 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v5s7x" podStartSLOduration=2.672994333 podStartE2EDuration="46.589467052s" podCreationTimestamp="2025-12-01 08:20:22 +0000 UTC" firstStartedPulling="2025-12-01 08:20:24.21930716 +0000 UTC m=+156.208365081" lastFinishedPulling="2025-12-01 08:21:08.135779879 +0000 UTC m=+200.124837800" observedRunningTime="2025-12-01 08:21:08.586560378 +0000 UTC m=+200.575618339" watchObservedRunningTime="2025-12-01 08:21:08.589467052 +0000 UTC m=+200.578524973" Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.775100 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.971796 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kube-api-access\") pod \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.972460 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kubelet-dir\") pod \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\" (UID: \"660aa0b3-48c4-4bfa-b712-f29e62fba0f4\") " Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.972746 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "660aa0b3-48c4-4bfa-b712-f29e62fba0f4" (UID: "660aa0b3-48c4-4bfa-b712-f29e62fba0f4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.972875 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:08 crc kubenswrapper[4744]: I1201 08:21:08.976401 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "660aa0b3-48c4-4bfa-b712-f29e62fba0f4" (UID: "660aa0b3-48c4-4bfa-b712-f29e62fba0f4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:21:09 crc kubenswrapper[4744]: I1201 08:21:09.073488 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/660aa0b3-48c4-4bfa-b712-f29e62fba0f4-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:09 crc kubenswrapper[4744]: I1201 08:21:09.571958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"660aa0b3-48c4-4bfa-b712-f29e62fba0f4","Type":"ContainerDied","Data":"3c64381cc18e32d0d93506211432b06da2e1ac16cf8af3a9c56d6095ac993193"} Dec 01 08:21:09 crc kubenswrapper[4744]: I1201 08:21:09.571998 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c64381cc18e32d0d93506211432b06da2e1ac16cf8af3a9c56d6095ac993193" Dec 01 08:21:09 crc kubenswrapper[4744]: I1201 08:21:09.572029 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:21:11 crc kubenswrapper[4744]: I1201 08:21:11.582128 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnzr5" event={"ID":"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b","Type":"ContainerStarted","Data":"c4248cbb3325cec506a610f3e7dba2afc7fb509400f322ab13a5395435c6efdf"} Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.589538 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md89c" event={"ID":"e9d0b50b-d228-4113-b88f-3ff873f6367e","Type":"ContainerStarted","Data":"81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043"} Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.594086 4744 generic.go:334] "Generic (PLEG): container finished" podID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerID="c4248cbb3325cec506a610f3e7dba2afc7fb509400f322ab13a5395435c6efdf" exitCode=0 Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.594151 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnzr5" event={"ID":"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b","Type":"ContainerDied","Data":"c4248cbb3325cec506a610f3e7dba2afc7fb509400f322ab13a5395435c6efdf"} Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.605652 4744 generic.go:334] "Generic (PLEG): container finished" podID="862392cd-33ed-4e07-8ea5-eef33e192765" containerID="c388bb3caf9de93279d4e066ba6b9a930f3321000920d2b96f1f37e7904ca741" exitCode=0 Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.605703 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rscr" event={"ID":"862392cd-33ed-4e07-8ea5-eef33e192765","Type":"ContainerDied","Data":"c388bb3caf9de93279d4e066ba6b9a930f3321000920d2b96f1f37e7904ca741"} Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.633875 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerID="85fc8e77989cd48773ebb98c7fb77218c0ea6e79718443322dd998438d034b06" exitCode=0 Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.633971 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvtt2" event={"ID":"d2e302f9-45f6-4e8e-9c08-984faf8732d5","Type":"ContainerDied","Data":"85fc8e77989cd48773ebb98c7fb77218c0ea6e79718443322dd998438d034b06"} Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.796277 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:21:12 crc kubenswrapper[4744]: E1201 08:21:12.796491 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="660aa0b3-48c4-4bfa-b712-f29e62fba0f4" containerName="pruner" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.796503 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="660aa0b3-48c4-4bfa-b712-f29e62fba0f4" containerName="pruner" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.796602 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="660aa0b3-48c4-4bfa-b712-f29e62fba0f4" containerName="pruner" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.796933 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.802922 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.803955 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.811010 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.850312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.850362 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e227ac63-afad-4e84-937d-98f39aa3d985-kube-api-access\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.850502 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-var-lock\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.933789 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.933861 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.951355 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.951429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e227ac63-afad-4e84-937d-98f39aa3d985-kube-api-access\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.951534 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-var-lock\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.951539 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.951611 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-var-lock\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:12 crc kubenswrapper[4744]: I1201 08:21:12.973023 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e227ac63-afad-4e84-937d-98f39aa3d985-kube-api-access\") pod \"installer-9-crc\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.066604 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.113237 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.264528 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:21:13 crc kubenswrapper[4744]: W1201 08:21:13.273455 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode227ac63_afad_4e84_937d_98f39aa3d985.slice/crio-b11ee3dc37c3f8647a2dc3f31aba1b931a479ad502908d947328fee891b00262 WatchSource:0}: Error finding container b11ee3dc37c3f8647a2dc3f31aba1b931a479ad502908d947328fee891b00262: Status 404 returned error can't find the container with id b11ee3dc37c3f8647a2dc3f31aba1b931a479ad502908d947328fee891b00262 Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.641596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvtt2" event={"ID":"d2e302f9-45f6-4e8e-9c08-984faf8732d5","Type":"ContainerStarted","Data":"b8ef7f269fd4b2371a93ad9795f22e8d16bf21806a0b2642c4d534003700caea"} Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.644022 4744 generic.go:334] "Generic (PLEG): container finished" podID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerID="81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043" exitCode=0 Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.644068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md89c" event={"ID":"e9d0b50b-d228-4113-b88f-3ff873f6367e","Type":"ContainerDied","Data":"81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043"} Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.646377 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e227ac63-afad-4e84-937d-98f39aa3d985","Type":"ContainerStarted","Data":"36f0af2aee920fd132a3bde0bc5bfa7c9dcab1f0fb0fd137a16ae2410de97f4a"} Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.646400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e227ac63-afad-4e84-937d-98f39aa3d985","Type":"ContainerStarted","Data":"b11ee3dc37c3f8647a2dc3f31aba1b931a479ad502908d947328fee891b00262"} Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.649500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnzr5" event={"ID":"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b","Type":"ContainerStarted","Data":"aad146934f9e5a08e12f5151fa75f724b41ac37c0e58dde86b827407ade64f2c"} Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.675131 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pvtt2" podStartSLOduration=2.500468848 podStartE2EDuration="51.675115063s" podCreationTimestamp="2025-12-01 08:20:22 +0000 UTC" firstStartedPulling="2025-12-01 08:20:24.197554933 +0000 UTC m=+156.186612854" lastFinishedPulling="2025-12-01 08:21:13.372201148 +0000 UTC m=+205.361259069" observedRunningTime="2025-12-01 08:21:13.664403181 +0000 UTC m=+205.653461102" watchObservedRunningTime="2025-12-01 08:21:13.675115063 +0000 UTC m=+205.664172984" Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.676253 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.676248556 podStartE2EDuration="1.676248556s" podCreationTimestamp="2025-12-01 08:21:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:21:13.674178386 +0000 UTC m=+205.663236307" watchObservedRunningTime="2025-12-01 08:21:13.676248556 +0000 UTC m=+205.665306477" Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.692785 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:21:13 crc kubenswrapper[4744]: I1201 08:21:13.695884 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pnzr5" podStartSLOduration=3.515982319 podStartE2EDuration="52.695872749s" podCreationTimestamp="2025-12-01 08:20:21 +0000 UTC" firstStartedPulling="2025-12-01 08:20:24.183789979 +0000 UTC m=+156.172847900" lastFinishedPulling="2025-12-01 08:21:13.363680409 +0000 UTC m=+205.352738330" observedRunningTime="2025-12-01 08:21:13.693648644 +0000 UTC m=+205.682706565" watchObservedRunningTime="2025-12-01 08:21:13.695872749 +0000 UTC m=+205.684930670" Dec 01 08:21:14 crc kubenswrapper[4744]: I1201 08:21:14.656266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rscr" event={"ID":"862392cd-33ed-4e07-8ea5-eef33e192765","Type":"ContainerStarted","Data":"48d15cab9941a94430f5fae348bbecd90592211f8107b9956c65be7e206cbbf3"} Dec 01 08:21:14 crc kubenswrapper[4744]: I1201 08:21:14.658036 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spj2t" event={"ID":"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482","Type":"ContainerStarted","Data":"46dbb9b04e473d90da1d310033c12b2578ed1b5360de678a89230beeccb1e679"} Dec 01 08:21:14 crc kubenswrapper[4744]: I1201 08:21:14.660246 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md89c" event={"ID":"e9d0b50b-d228-4113-b88f-3ff873f6367e","Type":"ContainerStarted","Data":"530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d"} Dec 01 08:21:14 crc kubenswrapper[4744]: I1201 08:21:14.662808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs9pb" event={"ID":"fe4049b3-153a-45d1-8847-a51f91053f77","Type":"ContainerStarted","Data":"d0c9a6e85783b095e2a3a26386797dd4499efa2bcd36990b19b3ebda43bda198"} Dec 01 08:21:14 crc kubenswrapper[4744]: I1201 08:21:14.696997 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6rscr" podStartSLOduration=1.737046645 podStartE2EDuration="50.6969797s" podCreationTimestamp="2025-12-01 08:20:24 +0000 UTC" firstStartedPulling="2025-12-01 08:20:25.225998618 +0000 UTC m=+157.215056539" lastFinishedPulling="2025-12-01 08:21:14.185931663 +0000 UTC m=+206.174989594" observedRunningTime="2025-12-01 08:21:14.696260069 +0000 UTC m=+206.685317990" watchObservedRunningTime="2025-12-01 08:21:14.6969797 +0000 UTC m=+206.686037621" Dec 01 08:21:14 crc kubenswrapper[4744]: I1201 08:21:14.715686 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-md89c" podStartSLOduration=2.837115248 podStartE2EDuration="52.715665985s" podCreationTimestamp="2025-12-01 08:20:22 +0000 UTC" firstStartedPulling="2025-12-01 08:20:24.214001592 +0000 UTC m=+156.203059513" lastFinishedPulling="2025-12-01 08:21:14.092552329 +0000 UTC m=+206.081610250" observedRunningTime="2025-12-01 08:21:14.709881196 +0000 UTC m=+206.698939117" watchObservedRunningTime="2025-12-01 08:21:14.715665985 +0000 UTC m=+206.704723906" Dec 01 08:21:15 crc kubenswrapper[4744]: I1201 08:21:15.672960 4744 generic.go:334] "Generic (PLEG): container finished" podID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerID="46dbb9b04e473d90da1d310033c12b2578ed1b5360de678a89230beeccb1e679" exitCode=0 Dec 01 08:21:15 crc kubenswrapper[4744]: I1201 08:21:15.673051 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spj2t" event={"ID":"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482","Type":"ContainerDied","Data":"46dbb9b04e473d90da1d310033c12b2578ed1b5360de678a89230beeccb1e679"} Dec 01 08:21:15 crc kubenswrapper[4744]: I1201 08:21:15.674573 4744 generic.go:334] "Generic (PLEG): container finished" podID="fe4049b3-153a-45d1-8847-a51f91053f77" containerID="d0c9a6e85783b095e2a3a26386797dd4499efa2bcd36990b19b3ebda43bda198" exitCode=0 Dec 01 08:21:15 crc kubenswrapper[4744]: I1201 08:21:15.674638 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs9pb" event={"ID":"fe4049b3-153a-45d1-8847-a51f91053f77","Type":"ContainerDied","Data":"d0c9a6e85783b095e2a3a26386797dd4499efa2bcd36990b19b3ebda43bda198"} Dec 01 08:21:15 crc kubenswrapper[4744]: I1201 08:21:15.676495 4744 generic.go:334] "Generic (PLEG): container finished" podID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerID="3fd04a2067fc34a5648e49b48f3184f84322bc853339bed94b8b6865b71e12f8" exitCode=0 Dec 01 08:21:15 crc kubenswrapper[4744]: I1201 08:21:15.676530 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngq48" event={"ID":"bd79b9ad-243e-470a-8c89-2a05be15e2ab","Type":"ContainerDied","Data":"3fd04a2067fc34a5648e49b48f3184f84322bc853339bed94b8b6865b71e12f8"} Dec 01 08:21:16 crc kubenswrapper[4744]: I1201 08:21:16.323501 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5s7x"] Dec 01 08:21:16 crc kubenswrapper[4744]: I1201 08:21:16.323687 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v5s7x" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="registry-server" containerID="cri-o://0a708ed95c06d19828bbe309d1645da3d9a2ad36bd51a83d6463fe42f52af9bc" gracePeriod=2 Dec 01 08:21:17 crc kubenswrapper[4744]: I1201 08:21:17.695360 4744 generic.go:334] "Generic (PLEG): container finished" podID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerID="0a708ed95c06d19828bbe309d1645da3d9a2ad36bd51a83d6463fe42f52af9bc" exitCode=0 Dec 01 08:21:17 crc kubenswrapper[4744]: I1201 08:21:17.695775 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5s7x" event={"ID":"2c6dc85b-81b3-42a1-92ea-93977c98dc1b","Type":"ContainerDied","Data":"0a708ed95c06d19828bbe309d1645da3d9a2ad36bd51a83d6463fe42f52af9bc"} Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.061655 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.122184 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cghb\" (UniqueName: \"kubernetes.io/projected/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-kube-api-access-5cghb\") pod \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.133616 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-kube-api-access-5cghb" (OuterVolumeSpecName: "kube-api-access-5cghb") pod "2c6dc85b-81b3-42a1-92ea-93977c98dc1b" (UID: "2c6dc85b-81b3-42a1-92ea-93977c98dc1b"). InnerVolumeSpecName "kube-api-access-5cghb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.226005 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-catalog-content\") pod \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.226093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-utilities\") pod \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\" (UID: \"2c6dc85b-81b3-42a1-92ea-93977c98dc1b\") " Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.226384 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cghb\" (UniqueName: \"kubernetes.io/projected/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-kube-api-access-5cghb\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.226806 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-utilities" (OuterVolumeSpecName: "utilities") pod "2c6dc85b-81b3-42a1-92ea-93977c98dc1b" (UID: "2c6dc85b-81b3-42a1-92ea-93977c98dc1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.327223 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.634178 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.634242 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.634288 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.634754 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.634860 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19" gracePeriod=600 Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.704955 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5s7x" event={"ID":"2c6dc85b-81b3-42a1-92ea-93977c98dc1b","Type":"ContainerDied","Data":"7c22a82ccfd773267c170ecf9f30823bea00be206f45011a21893a5c5bc89852"} Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.705021 4744 scope.go:117] "RemoveContainer" containerID="0a708ed95c06d19828bbe309d1645da3d9a2ad36bd51a83d6463fe42f52af9bc" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.705054 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5s7x" Dec 01 08:21:18 crc kubenswrapper[4744]: I1201 08:21:18.958448 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c6dc85b-81b3-42a1-92ea-93977c98dc1b" (UID: "2c6dc85b-81b3-42a1-92ea-93977c98dc1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:19 crc kubenswrapper[4744]: I1201 08:21:19.037228 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c6dc85b-81b3-42a1-92ea-93977c98dc1b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:19 crc kubenswrapper[4744]: I1201 08:21:19.059986 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5s7x"] Dec 01 08:21:19 crc kubenswrapper[4744]: I1201 08:21:19.060061 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v5s7x"] Dec 01 08:21:19 crc kubenswrapper[4744]: I1201 08:21:19.712782 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19" exitCode=0 Dec 01 08:21:19 crc kubenswrapper[4744]: I1201 08:21:19.712876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19"} Dec 01 08:21:20 crc kubenswrapper[4744]: I1201 08:21:20.006635 4744 scope.go:117] "RemoveContainer" containerID="16ecd5a22c1907203d773d276bf2b1da861abea3bb645d063710448d6264b066" Dec 01 08:21:20 crc kubenswrapper[4744]: I1201 08:21:20.295785 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" path="/var/lib/kubelet/pods/2c6dc85b-81b3-42a1-92ea-93977c98dc1b/volumes" Dec 01 08:21:20 crc kubenswrapper[4744]: I1201 08:21:20.953986 4744 scope.go:117] "RemoveContainer" containerID="a1eb48880392d3e966454f7be898e6b98f8c4bcdea3113dc2848febc78cb18e7" Dec 01 08:21:21 crc kubenswrapper[4744]: I1201 08:21:21.743163 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs9pb" event={"ID":"fe4049b3-153a-45d1-8847-a51f91053f77","Type":"ContainerStarted","Data":"6325a8ee70f509963ff6c9d9dcadcbe620b52d357e95b1ce0229e2e4cae1aa45"} Dec 01 08:21:21 crc kubenswrapper[4744]: I1201 08:21:21.752239 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spj2t" event={"ID":"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482","Type":"ContainerStarted","Data":"a46386e833d7931c9c33a038db763cb24dae3791ecb0bb3848d2f9235b4205bd"} Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.335431 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.335827 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.394461 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.538470 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.538820 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.597574 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.674424 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.675266 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.712703 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.759263 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"4b444480d1eb4ea273efbce3cf9243860be3f4dff96451c41a797893d150f669"} Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.794757 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.799667 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:21:22 crc kubenswrapper[4744]: I1201 08:21:22.825955 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:21:23 crc kubenswrapper[4744]: I1201 08:21:23.787433 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-spj2t" podStartSLOduration=5.137815415 podStartE2EDuration="58.787404622s" podCreationTimestamp="2025-12-01 08:20:25 +0000 UTC" firstStartedPulling="2025-12-01 08:20:27.304666027 +0000 UTC m=+159.293723948" lastFinishedPulling="2025-12-01 08:21:20.954255234 +0000 UTC m=+212.943313155" observedRunningTime="2025-12-01 08:21:23.78526756 +0000 UTC m=+215.774325481" watchObservedRunningTime="2025-12-01 08:21:23.787404622 +0000 UTC m=+215.776462543" Dec 01 08:21:23 crc kubenswrapper[4744]: I1201 08:21:23.821469 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gs9pb" podStartSLOduration=3.907921289 podStartE2EDuration="58.821439155s" podCreationTimestamp="2025-12-01 08:20:25 +0000 UTC" firstStartedPulling="2025-12-01 08:20:26.287387243 +0000 UTC m=+158.276445164" lastFinishedPulling="2025-12-01 08:21:21.200905109 +0000 UTC m=+213.189963030" observedRunningTime="2025-12-01 08:21:23.820469827 +0000 UTC m=+215.809527748" watchObservedRunningTime="2025-12-01 08:21:23.821439155 +0000 UTC m=+215.810497076" Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.434777 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.435172 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.541786 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.723512 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-md89c"] Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.771203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngq48" event={"ID":"bd79b9ad-243e-470a-8c89-2a05be15e2ab","Type":"ContainerStarted","Data":"3147241bc56e044b9b13874068c95d804b042fe21696736fb27814ecc755a44e"} Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.796770 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ngq48" podStartSLOduration=3.070553331 podStartE2EDuration="1m0.796753094s" podCreationTimestamp="2025-12-01 08:20:24 +0000 UTC" firstStartedPulling="2025-12-01 08:20:26.235477003 +0000 UTC m=+158.224534924" lastFinishedPulling="2025-12-01 08:21:23.961676766 +0000 UTC m=+215.950734687" observedRunningTime="2025-12-01 08:21:24.791561772 +0000 UTC m=+216.780619703" watchObservedRunningTime="2025-12-01 08:21:24.796753094 +0000 UTC m=+216.785811015" Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.818934 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.841044 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:21:24 crc kubenswrapper[4744]: I1201 08:21:24.841094 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:21:25 crc kubenswrapper[4744]: I1201 08:21:25.447936 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:21:25 crc kubenswrapper[4744]: I1201 08:21:25.448334 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:21:25 crc kubenswrapper[4744]: I1201 08:21:25.777475 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-md89c" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="registry-server" containerID="cri-o://530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d" gracePeriod=2 Dec 01 08:21:25 crc kubenswrapper[4744]: I1201 08:21:25.855640 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:21:25 crc kubenswrapper[4744]: I1201 08:21:25.855724 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:21:25 crc kubenswrapper[4744]: I1201 08:21:25.878760 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-ngq48" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="registry-server" probeResult="failure" output=< Dec 01 08:21:25 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 08:21:25 crc kubenswrapper[4744]: > Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.491938 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gs9pb" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="registry-server" probeResult="failure" output=< Dec 01 08:21:26 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 08:21:26 crc kubenswrapper[4744]: > Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.682675 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.747504 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-utilities\") pod \"e9d0b50b-d228-4113-b88f-3ff873f6367e\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.747889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqsdl\" (UniqueName: \"kubernetes.io/projected/e9d0b50b-d228-4113-b88f-3ff873f6367e-kube-api-access-mqsdl\") pod \"e9d0b50b-d228-4113-b88f-3ff873f6367e\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.747927 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-catalog-content\") pod \"e9d0b50b-d228-4113-b88f-3ff873f6367e\" (UID: \"e9d0b50b-d228-4113-b88f-3ff873f6367e\") " Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.749818 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-utilities" (OuterVolumeSpecName: "utilities") pod "e9d0b50b-d228-4113-b88f-3ff873f6367e" (UID: "e9d0b50b-d228-4113-b88f-3ff873f6367e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.753225 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d0b50b-d228-4113-b88f-3ff873f6367e-kube-api-access-mqsdl" (OuterVolumeSpecName: "kube-api-access-mqsdl") pod "e9d0b50b-d228-4113-b88f-3ff873f6367e" (UID: "e9d0b50b-d228-4113-b88f-3ff873f6367e"). InnerVolumeSpecName "kube-api-access-mqsdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.783796 4744 generic.go:334] "Generic (PLEG): container finished" podID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerID="530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d" exitCode=0 Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.783855 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md89c" event={"ID":"e9d0b50b-d228-4113-b88f-3ff873f6367e","Type":"ContainerDied","Data":"530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d"} Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.783900 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-md89c" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.783931 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-md89c" event={"ID":"e9d0b50b-d228-4113-b88f-3ff873f6367e","Type":"ContainerDied","Data":"8b77e42c89881ff17a162f8ba771f893e1bf5917e970c08223dca23056f093a3"} Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.783965 4744 scope.go:117] "RemoveContainer" containerID="530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.793783 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9d0b50b-d228-4113-b88f-3ff873f6367e" (UID: "e9d0b50b-d228-4113-b88f-3ff873f6367e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.808941 4744 scope.go:117] "RemoveContainer" containerID="81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.823493 4744 scope.go:117] "RemoveContainer" containerID="04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.842250 4744 scope.go:117] "RemoveContainer" containerID="530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d" Dec 01 08:21:26 crc kubenswrapper[4744]: E1201 08:21:26.842971 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d\": container with ID starting with 530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d not found: ID does not exist" containerID="530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.843018 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d"} err="failed to get container status \"530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d\": rpc error: code = NotFound desc = could not find container \"530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d\": container with ID starting with 530d187276fdffce285bd8d96ed1eef0a468129b40957ff6d9f8b7ff08a9069d not found: ID does not exist" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.843048 4744 scope.go:117] "RemoveContainer" containerID="81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043" Dec 01 08:21:26 crc kubenswrapper[4744]: E1201 08:21:26.843973 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043\": container with ID starting with 81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043 not found: ID does not exist" containerID="81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.844022 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043"} err="failed to get container status \"81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043\": rpc error: code = NotFound desc = could not find container \"81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043\": container with ID starting with 81c2034891b8d9bb9becb5fd30c250ce41e672160e2cececc9880aad55e1b043 not found: ID does not exist" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.844056 4744 scope.go:117] "RemoveContainer" containerID="04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b" Dec 01 08:21:26 crc kubenswrapper[4744]: E1201 08:21:26.844658 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b\": container with ID starting with 04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b not found: ID does not exist" containerID="04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.844697 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b"} err="failed to get container status \"04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b\": rpc error: code = NotFound desc = could not find container \"04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b\": container with ID starting with 04e8e0d551df9a9ed258c250b3990e7622ac5b9bd1e0a14dd1ca035d1cd3830b not found: ID does not exist" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.849556 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqsdl\" (UniqueName: \"kubernetes.io/projected/e9d0b50b-d228-4113-b88f-3ff873f6367e-kube-api-access-mqsdl\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.849588 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.849604 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9d0b50b-d228-4113-b88f-3ff873f6367e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:26 crc kubenswrapper[4744]: I1201 08:21:26.923941 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-spj2t" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="registry-server" probeResult="failure" output=< Dec 01 08:21:26 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 08:21:26 crc kubenswrapper[4744]: > Dec 01 08:21:27 crc kubenswrapper[4744]: I1201 08:21:27.110161 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-md89c"] Dec 01 08:21:27 crc kubenswrapper[4744]: I1201 08:21:27.125308 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-md89c"] Dec 01 08:21:28 crc kubenswrapper[4744]: I1201 08:21:28.311003 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" path="/var/lib/kubelet/pods/e9d0b50b-d228-4113-b88f-3ff873f6367e/volumes" Dec 01 08:21:34 crc kubenswrapper[4744]: I1201 08:21:34.534582 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pflgv"] Dec 01 08:21:34 crc kubenswrapper[4744]: I1201 08:21:34.876737 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:21:34 crc kubenswrapper[4744]: I1201 08:21:34.910055 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:21:35 crc kubenswrapper[4744]: I1201 08:21:35.496445 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:21:35 crc kubenswrapper[4744]: I1201 08:21:35.538114 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:21:35 crc kubenswrapper[4744]: I1201 08:21:35.906933 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:21:35 crc kubenswrapper[4744]: I1201 08:21:35.953091 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:21:36 crc kubenswrapper[4744]: I1201 08:21:36.363759 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngq48"] Dec 01 08:21:36 crc kubenswrapper[4744]: I1201 08:21:36.843581 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ngq48" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="registry-server" containerID="cri-o://3147241bc56e044b9b13874068c95d804b042fe21696736fb27814ecc755a44e" gracePeriod=2 Dec 01 08:21:37 crc kubenswrapper[4744]: I1201 08:21:37.762615 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-spj2t"] Dec 01 08:21:37 crc kubenswrapper[4744]: I1201 08:21:37.848673 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-spj2t" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="registry-server" containerID="cri-o://a46386e833d7931c9c33a038db763cb24dae3791ecb0bb3848d2f9235b4205bd" gracePeriod=2 Dec 01 08:21:38 crc kubenswrapper[4744]: I1201 08:21:38.864459 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngq48" event={"ID":"bd79b9ad-243e-470a-8c89-2a05be15e2ab","Type":"ContainerDied","Data":"3147241bc56e044b9b13874068c95d804b042fe21696736fb27814ecc755a44e"} Dec 01 08:21:38 crc kubenswrapper[4744]: I1201 08:21:38.864481 4744 generic.go:334] "Generic (PLEG): container finished" podID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerID="3147241bc56e044b9b13874068c95d804b042fe21696736fb27814ecc755a44e" exitCode=0 Dec 01 08:21:38 crc kubenswrapper[4744]: I1201 08:21:38.867621 4744 generic.go:334] "Generic (PLEG): container finished" podID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerID="a46386e833d7931c9c33a038db763cb24dae3791ecb0bb3848d2f9235b4205bd" exitCode=0 Dec 01 08:21:38 crc kubenswrapper[4744]: I1201 08:21:38.867673 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spj2t" event={"ID":"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482","Type":"ContainerDied","Data":"a46386e833d7931c9c33a038db763cb24dae3791ecb0bb3848d2f9235b4205bd"} Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.160969 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.166765 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.279541 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2vxb\" (UniqueName: \"kubernetes.io/projected/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-kube-api-access-h2vxb\") pod \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.279628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-catalog-content\") pod \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.279687 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-catalog-content\") pod \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.279761 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-utilities\") pod \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.279851 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k7k9\" (UniqueName: \"kubernetes.io/projected/bd79b9ad-243e-470a-8c89-2a05be15e2ab-kube-api-access-9k7k9\") pod \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\" (UID: \"bd79b9ad-243e-470a-8c89-2a05be15e2ab\") " Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.279907 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-utilities\") pod \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\" (UID: \"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482\") " Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.281541 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-utilities" (OuterVolumeSpecName: "utilities") pod "bd79b9ad-243e-470a-8c89-2a05be15e2ab" (UID: "bd79b9ad-243e-470a-8c89-2a05be15e2ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.285497 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-utilities" (OuterVolumeSpecName: "utilities") pod "945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" (UID: "945f44ee-e2a3-4a1a-a6c0-4cd6b107c482"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.286040 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd79b9ad-243e-470a-8c89-2a05be15e2ab-kube-api-access-9k7k9" (OuterVolumeSpecName: "kube-api-access-9k7k9") pod "bd79b9ad-243e-470a-8c89-2a05be15e2ab" (UID: "bd79b9ad-243e-470a-8c89-2a05be15e2ab"). InnerVolumeSpecName "kube-api-access-9k7k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.287643 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-kube-api-access-h2vxb" (OuterVolumeSpecName: "kube-api-access-h2vxb") pod "945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" (UID: "945f44ee-e2a3-4a1a-a6c0-4cd6b107c482"). InnerVolumeSpecName "kube-api-access-h2vxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.305246 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd79b9ad-243e-470a-8c89-2a05be15e2ab" (UID: "bd79b9ad-243e-470a-8c89-2a05be15e2ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.382561 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.382617 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd79b9ad-243e-470a-8c89-2a05be15e2ab-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.382636 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k7k9\" (UniqueName: \"kubernetes.io/projected/bd79b9ad-243e-470a-8c89-2a05be15e2ab-kube-api-access-9k7k9\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.382660 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.382678 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2vxb\" (UniqueName: \"kubernetes.io/projected/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-kube-api-access-h2vxb\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.414290 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" (UID: "945f44ee-e2a3-4a1a-a6c0-4cd6b107c482"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.483291 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.875919 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngq48" event={"ID":"bd79b9ad-243e-470a-8c89-2a05be15e2ab","Type":"ContainerDied","Data":"0c69f76fdb27a92c1d338bbbc19473e41b7bdd8377e188b747719e41880556e1"} Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.876960 4744 scope.go:117] "RemoveContainer" containerID="3147241bc56e044b9b13874068c95d804b042fe21696736fb27814ecc755a44e" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.876034 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngq48" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.879929 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spj2t" event={"ID":"945f44ee-e2a3-4a1a-a6c0-4cd6b107c482","Type":"ContainerDied","Data":"14d7586e56ff6c3f96bc30c65e6e859101f631914850677ef292d6b1383d800d"} Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.880046 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spj2t" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.901911 4744 scope.go:117] "RemoveContainer" containerID="3fd04a2067fc34a5648e49b48f3184f84322bc853339bed94b8b6865b71e12f8" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.911493 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngq48"] Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.919343 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngq48"] Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.925816 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-spj2t"] Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.926775 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-spj2t"] Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.936151 4744 scope.go:117] "RemoveContainer" containerID="8713cb16920f45df71c3eca2e7514466eaf261e6406a91f8b70cf8affb6421dd" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.947060 4744 scope.go:117] "RemoveContainer" containerID="a46386e833d7931c9c33a038db763cb24dae3791ecb0bb3848d2f9235b4205bd" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.961774 4744 scope.go:117] "RemoveContainer" containerID="46dbb9b04e473d90da1d310033c12b2578ed1b5360de678a89230beeccb1e679" Dec 01 08:21:39 crc kubenswrapper[4744]: I1201 08:21:39.976322 4744 scope.go:117] "RemoveContainer" containerID="5937f44f65ae6cbd2308697dcd6cb23c925c6f62f7eed4b8a00bbe20231d00f1" Dec 01 08:21:40 crc kubenswrapper[4744]: I1201 08:21:40.298349 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" path="/var/lib/kubelet/pods/945f44ee-e2a3-4a1a-a6c0-4cd6b107c482/volumes" Dec 01 08:21:40 crc kubenswrapper[4744]: I1201 08:21:40.298961 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" path="/var/lib/kubelet/pods/bd79b9ad-243e-470a-8c89-2a05be15e2ab/volumes" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.185205 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186382 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186434 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186475 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186489 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186501 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186516 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186528 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186547 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186558 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186577 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186588 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186609 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186621 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186636 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186648 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186667 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186679 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186692 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186704 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="extract-content" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186717 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186729 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.186746 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186760 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="extract-utilities" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186958 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="945f44ee-e2a3-4a1a-a6c0-4cd6b107c482" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.186993 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d0b50b-d228-4113-b88f-3ff873f6367e" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.187012 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd79b9ad-243e-470a-8c89-2a05be15e2ab" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.187026 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c6dc85b-81b3-42a1-92ea-93977c98dc1b" containerName="registry-server" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.187559 4744 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.187800 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.188055 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918" gracePeriod=15 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.188102 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778" gracePeriod=15 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.188073 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0" gracePeriod=15 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.188174 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab" gracePeriod=15 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.188880 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.188174 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97" gracePeriod=15 Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.189183 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189209 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.189237 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189254 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.189271 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189286 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.189310 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189327 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.189349 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189365 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.189390 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189470 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:21:51 crc kubenswrapper[4744]: E1201 08:21:51.189506 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189522 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189732 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189754 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189777 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189796 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189824 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.189846 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.234801 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.235178 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.235254 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.235305 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.235338 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.235376 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.235460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.235525 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341009 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341067 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341120 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341163 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341198 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341304 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341376 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341426 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341468 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341493 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341548 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.341572 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.961210 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.963527 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.965115 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0" exitCode=0 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.965167 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97" exitCode=0 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.965185 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778" exitCode=0 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.965199 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab" exitCode=2 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.965263 4744 scope.go:117] "RemoveContainer" containerID="d0a16f69fbf39a8c4756b00d07a8cee3a972a5c5d163c322f1922faaed9dd902" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.968384 4744 generic.go:334] "Generic (PLEG): container finished" podID="e227ac63-afad-4e84-937d-98f39aa3d985" containerID="36f0af2aee920fd132a3bde0bc5bfa7c9dcab1f0fb0fd137a16ae2410de97f4a" exitCode=0 Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.968488 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e227ac63-afad-4e84-937d-98f39aa3d985","Type":"ContainerDied","Data":"36f0af2aee920fd132a3bde0bc5bfa7c9dcab1f0fb0fd137a16ae2410de97f4a"} Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.969923 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:51 crc kubenswrapper[4744]: I1201 08:21:51.971108 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:52 crc kubenswrapper[4744]: I1201 08:21:52.977931 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.266482 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.267362 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.368671 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-kubelet-dir\") pod \"e227ac63-afad-4e84-937d-98f39aa3d985\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.368889 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e227ac63-afad-4e84-937d-98f39aa3d985" (UID: "e227ac63-afad-4e84-937d-98f39aa3d985"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.369170 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e227ac63-afad-4e84-937d-98f39aa3d985-kube-api-access\") pod \"e227ac63-afad-4e84-937d-98f39aa3d985\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.369206 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-var-lock\") pod \"e227ac63-afad-4e84-937d-98f39aa3d985\" (UID: \"e227ac63-afad-4e84-937d-98f39aa3d985\") " Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.369468 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.370311 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-var-lock" (OuterVolumeSpecName: "var-lock") pod "e227ac63-afad-4e84-937d-98f39aa3d985" (UID: "e227ac63-afad-4e84-937d-98f39aa3d985"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.379028 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e227ac63-afad-4e84-937d-98f39aa3d985-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e227ac63-afad-4e84-937d-98f39aa3d985" (UID: "e227ac63-afad-4e84-937d-98f39aa3d985"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.470660 4744 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e227ac63-afad-4e84-937d-98f39aa3d985-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.470694 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e227ac63-afad-4e84-937d-98f39aa3d985-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.584138 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.585291 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.585868 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.586089 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.774655 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.774736 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.774758 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.774836 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.774859 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.774908 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.775033 4744 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.775055 4744 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.775074 4744 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:21:53 crc kubenswrapper[4744]: E1201 08:21:53.895942 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: E1201 08:21:53.896559 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: E1201 08:21:53.897159 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: E1201 08:21:53.897754 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: E1201 08:21:53.898175 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:53 crc kubenswrapper[4744]: I1201 08:21:53.898220 4744 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 08:21:53 crc kubenswrapper[4744]: E1201 08:21:53.898490 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="200ms" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.002929 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.003901 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918" exitCode=0 Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.003987 4744 scope.go:117] "RemoveContainer" containerID="a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.004010 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.006043 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e227ac63-afad-4e84-937d-98f39aa3d985","Type":"ContainerDied","Data":"b11ee3dc37c3f8647a2dc3f31aba1b931a479ad502908d947328fee891b00262"} Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.006064 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b11ee3dc37c3f8647a2dc3f31aba1b931a479ad502908d947328fee891b00262" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.006109 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.026328 4744 scope.go:117] "RemoveContainer" containerID="a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.033878 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.034191 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.034526 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.034761 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.042928 4744 scope.go:117] "RemoveContainer" containerID="84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.058928 4744 scope.go:117] "RemoveContainer" containerID="b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.075985 4744 scope.go:117] "RemoveContainer" containerID="fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.099283 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="400ms" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.101050 4744 scope.go:117] "RemoveContainer" containerID="d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.118279 4744 scope.go:117] "RemoveContainer" containerID="a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.118652 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\": container with ID starting with a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0 not found: ID does not exist" containerID="a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.118686 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0"} err="failed to get container status \"a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\": rpc error: code = NotFound desc = could not find container \"a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0\": container with ID starting with a54b1ca70676adb289ad436adb0bc0d9bdd6000ca097862e4a364d53f10ed7d0 not found: ID does not exist" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.118708 4744 scope.go:117] "RemoveContainer" containerID="a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.118950 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\": container with ID starting with a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97 not found: ID does not exist" containerID="a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.118977 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97"} err="failed to get container status \"a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\": rpc error: code = NotFound desc = could not find container \"a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97\": container with ID starting with a204eb642202fef92ef71d119a239183ec43ee1acbea83261f9a16fe69700b97 not found: ID does not exist" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.118991 4744 scope.go:117] "RemoveContainer" containerID="84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.119218 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\": container with ID starting with 84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778 not found: ID does not exist" containerID="84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.119239 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778"} err="failed to get container status \"84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\": rpc error: code = NotFound desc = could not find container \"84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778\": container with ID starting with 84fe73209d4dcbb5421242a6442c5d36233e0184723908230d94599fae029778 not found: ID does not exist" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.119253 4744 scope.go:117] "RemoveContainer" containerID="b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.119541 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\": container with ID starting with b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab not found: ID does not exist" containerID="b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.119571 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab"} err="failed to get container status \"b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\": rpc error: code = NotFound desc = could not find container \"b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab\": container with ID starting with b1fae710eff422db1acc9fdbdc234994eb8ca9c066c58dfeffbd19a7aa9b7aab not found: ID does not exist" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.119587 4744 scope.go:117] "RemoveContainer" containerID="fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.119793 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\": container with ID starting with fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918 not found: ID does not exist" containerID="fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.119818 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918"} err="failed to get container status \"fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\": rpc error: code = NotFound desc = could not find container \"fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918\": container with ID starting with fa2f7f8cf91c00c80bbe1e4139344813e39f36365df774000b923bc425fb5918 not found: ID does not exist" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.119871 4744 scope.go:117] "RemoveContainer" containerID="d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.120633 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\": container with ID starting with d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc not found: ID does not exist" containerID="d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.120659 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc"} err="failed to get container status \"d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\": rpc error: code = NotFound desc = could not find container \"d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc\": container with ID starting with d6d828c63e7331342f477f01fa783cfdcb568b0b92a851a0fe61780068fbfacc not found: ID does not exist" Dec 01 08:21:54 crc kubenswrapper[4744]: I1201 08:21:54.291435 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 08:21:54 crc kubenswrapper[4744]: E1201 08:21:54.502288 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="800ms" Dec 01 08:21:55 crc kubenswrapper[4744]: E1201 08:21:55.304522 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="1.6s" Dec 01 08:21:56 crc kubenswrapper[4744]: E1201 08:21:56.247344 4744 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:56 crc kubenswrapper[4744]: I1201 08:21:56.248021 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:56 crc kubenswrapper[4744]: E1201 08:21:56.281279 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.39:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d09b853996297 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:21:56.280296087 +0000 UTC m=+248.269354058,LastTimestamp:2025-12-01 08:21:56.280296087 +0000 UTC m=+248.269354058,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:21:56 crc kubenswrapper[4744]: E1201 08:21:56.906333 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="3.2s" Dec 01 08:21:57 crc kubenswrapper[4744]: I1201 08:21:57.023384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09"} Dec 01 08:21:57 crc kubenswrapper[4744]: I1201 08:21:57.023455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cb2ccc077af068609c414342ae2bfff3cde2f39212cecbfcbc2d06c46fe4cf80"} Dec 01 08:21:57 crc kubenswrapper[4744]: I1201 08:21:57.024021 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:57 crc kubenswrapper[4744]: E1201 08:21:57.024068 4744 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:21:58 crc kubenswrapper[4744]: I1201 08:21:58.289392 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:21:59 crc kubenswrapper[4744]: I1201 08:21:59.573673 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" containerName="oauth-openshift" containerID="cri-o://2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5" gracePeriod=15 Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.044721 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.045566 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.045909 4744 status_manager.go:851] "Failed to get status for pod" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pflgv\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.046010 4744 generic.go:334] "Generic (PLEG): container finished" podID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" containerID="2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5" exitCode=0 Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.046052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" event={"ID":"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61","Type":"ContainerDied","Data":"2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5"} Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.046086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" event={"ID":"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61","Type":"ContainerDied","Data":"332b06aaabb88f879f1d94955a195195faf99bc8de6f62095b6aad17e46a3a1f"} Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.046110 4744 scope.go:117] "RemoveContainer" containerID="2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.061276 4744 scope.go:117] "RemoveContainer" containerID="2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5" Dec 01 08:22:00 crc kubenswrapper[4744]: E1201 08:22:00.061782 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5\": container with ID starting with 2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5 not found: ID does not exist" containerID="2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.061832 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5"} err="failed to get container status \"2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5\": rpc error: code = NotFound desc = could not find container \"2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5\": container with ID starting with 2126697e4e75c94ccd62ae9c4f31b33f5b8ed5f22c03eda2d3992811a022afa5 not found: ID does not exist" Dec 01 08:22:00 crc kubenswrapper[4744]: E1201 08:22:00.108136 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.39:6443: connect: connection refused" interval="6.4s" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.171218 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-policies\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.171289 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-dir\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.171371 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.171484 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-session\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.171900 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.171513 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-login\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172358 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-cliconfig\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172388 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-error\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172454 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-idp-0-file-data\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172484 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-trusted-ca-bundle\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172521 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-router-certs\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-serving-cert\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172571 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-service-ca\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172595 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-ocp-branding-template\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172615 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vclbj\" (UniqueName: \"kubernetes.io/projected/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-kube-api-access-vclbj\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172652 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-provider-selection\") pod \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\" (UID: \"34c76ecb-9f6d-4c8e-b147-5ac1b981cb61\") " Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172948 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.172964 4744 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.174275 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.174638 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.174803 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.177934 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.178899 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.180035 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.181119 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.181705 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.181835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-kube-api-access-vclbj" (OuterVolumeSpecName: "kube-api-access-vclbj") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "kube-api-access-vclbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.182723 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.183067 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.184031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" (UID: "34c76ecb-9f6d-4c8e-b147-5ac1b981cb61"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274306 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274376 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274429 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274461 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274486 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vclbj\" (UniqueName: \"kubernetes.io/projected/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-kube-api-access-vclbj\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274513 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274543 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274567 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274591 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274615 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.274638 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:00 crc kubenswrapper[4744]: I1201 08:22:00.275512 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:01 crc kubenswrapper[4744]: I1201 08:22:01.051864 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" Dec 01 08:22:01 crc kubenswrapper[4744]: I1201 08:22:01.052518 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:01 crc kubenswrapper[4744]: I1201 08:22:01.052828 4744 status_manager.go:851] "Failed to get status for pod" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pflgv\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:01 crc kubenswrapper[4744]: I1201 08:22:01.055343 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:01 crc kubenswrapper[4744]: I1201 08:22:01.055622 4744 status_manager.go:851] "Failed to get status for pod" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pflgv\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:03 crc kubenswrapper[4744]: E1201 08:22:03.312330 4744 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" volumeName="registry-storage" Dec 01 08:22:04 crc kubenswrapper[4744]: I1201 08:22:04.285327 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:04 crc kubenswrapper[4744]: I1201 08:22:04.288793 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:04 crc kubenswrapper[4744]: I1201 08:22:04.289807 4744 status_manager.go:851] "Failed to get status for pod" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pflgv\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:04 crc kubenswrapper[4744]: I1201 08:22:04.305944 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:04 crc kubenswrapper[4744]: I1201 08:22:04.305992 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:04 crc kubenswrapper[4744]: E1201 08:22:04.306665 4744 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:04 crc kubenswrapper[4744]: I1201 08:22:04.307290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.078951 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.079285 4744 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c" exitCode=1 Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.079382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c"} Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.080218 4744 scope.go:117] "RemoveContainer" containerID="af76c79b0e610cc298e5c55b9c1cf58ecb504273c7258fa874a763b192a31b7c" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.080294 4744 status_manager.go:851] "Failed to get status for pod" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pflgv\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.080822 4744 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.081185 4744 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9b53ea0ce77865b69fac0796f2657ad652b2dcc9a51f9f4a62d07b8a75e03d72" exitCode=0 Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.081181 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.081218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9b53ea0ce77865b69fac0796f2657ad652b2dcc9a51f9f4a62d07b8a75e03d72"} Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.081250 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cf88c6035ce50258a9cf6b754bcb7fd09ed3e6db8a9dfb1ba2676fac639342e8"} Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.081538 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.081553 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.081959 4744 status_manager.go:851] "Failed to get status for pod" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" pod="openshift-authentication/oauth-openshift-558db77b4-pflgv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-pflgv\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:05 crc kubenswrapper[4744]: E1201 08:22:05.081959 4744 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.082229 4744 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:05 crc kubenswrapper[4744]: I1201 08:22:05.082466 4744 status_manager.go:851] "Failed to get status for pod" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.39:6443: connect: connection refused" Dec 01 08:22:06 crc kubenswrapper[4744]: I1201 08:22:06.096742 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 08:22:06 crc kubenswrapper[4744]: I1201 08:22:06.097038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c879adaf6001f92afbdfa678f565cdcf3bf1b6d0a17953cab8178ca75d407d77"} Dec 01 08:22:06 crc kubenswrapper[4744]: I1201 08:22:06.102272 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"86d61e1a9bd176df8d9040b95c0c3fe756f622cb741d4823ff02007c1c08afc0"} Dec 01 08:22:06 crc kubenswrapper[4744]: I1201 08:22:06.102304 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"99d70d1bd8871f26fecde03fe1cdff2f5213b16be1627edc9c2f26e96fe3bfe2"} Dec 01 08:22:06 crc kubenswrapper[4744]: I1201 08:22:06.102317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"79dcf79ea5410181098e29b990543045f15107c5e1c5cfb332f1e1ebd27d0d64"} Dec 01 08:22:07 crc kubenswrapper[4744]: I1201 08:22:07.079740 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:22:07 crc kubenswrapper[4744]: I1201 08:22:07.079969 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 08:22:07 crc kubenswrapper[4744]: I1201 08:22:07.080049 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 08:22:07 crc kubenswrapper[4744]: I1201 08:22:07.112353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"09ae7e510a7758b04004ad9318f5c8a4ab104ff81fb2725a3893c6d48f360ed4"} Dec 01 08:22:07 crc kubenswrapper[4744]: I1201 08:22:07.113092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8e3ed7eb06c52bfce8868eae3f29deefb34106e13f0ae7d397f3dda581f01fae"} Dec 01 08:22:07 crc kubenswrapper[4744]: I1201 08:22:07.112774 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:07 crc kubenswrapper[4744]: I1201 08:22:07.113322 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:09 crc kubenswrapper[4744]: I1201 08:22:09.308467 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:09 crc kubenswrapper[4744]: I1201 08:22:09.308547 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:09 crc kubenswrapper[4744]: I1201 08:22:09.314996 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:11 crc kubenswrapper[4744]: I1201 08:22:11.982542 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:22:12 crc kubenswrapper[4744]: I1201 08:22:12.123223 4744 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:13 crc kubenswrapper[4744]: I1201 08:22:13.155069 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:13 crc kubenswrapper[4744]: I1201 08:22:13.155083 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:13 crc kubenswrapper[4744]: I1201 08:22:13.155741 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:13 crc kubenswrapper[4744]: I1201 08:22:13.162244 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:13 crc kubenswrapper[4744]: I1201 08:22:13.164922 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0f1f8948-47e8-4807-99a7-13f43a0c9ba1" Dec 01 08:22:14 crc kubenswrapper[4744]: I1201 08:22:14.162326 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:14 crc kubenswrapper[4744]: I1201 08:22:14.162371 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:15 crc kubenswrapper[4744]: I1201 08:22:15.169396 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:15 crc kubenswrapper[4744]: I1201 08:22:15.169796 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4ceb4de2-f0af-4141-a088-28f6a7525832" Dec 01 08:22:17 crc kubenswrapper[4744]: I1201 08:22:17.086898 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:22:17 crc kubenswrapper[4744]: I1201 08:22:17.096712 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:22:18 crc kubenswrapper[4744]: I1201 08:22:18.306265 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0f1f8948-47e8-4807-99a7-13f43a0c9ba1" Dec 01 08:22:22 crc kubenswrapper[4744]: I1201 08:22:22.664739 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.216336 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.391674 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.422486 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.480346 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.491388 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.667066 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.718489 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.953819 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 08:22:23 crc kubenswrapper[4744]: I1201 08:22:23.985587 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.280613 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.288598 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.314531 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.315615 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.334209 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.418038 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.451810 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.474634 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.578702 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.612105 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.720678 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.790603 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 08:22:24 crc kubenswrapper[4744]: I1201 08:22:24.943846 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.000020 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.292102 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.351776 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.461648 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.589118 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.808152 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.838751 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 08:22:25 crc kubenswrapper[4744]: I1201 08:22:25.843863 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.079134 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.127182 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.142165 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.158457 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.196222 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.296029 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.321792 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.683167 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.757989 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.784625 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.796652 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.945531 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 08:22:26 crc kubenswrapper[4744]: I1201 08:22:26.959676 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.038054 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.062692 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.107471 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.114235 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.134392 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.174239 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.428613 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.458354 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.489208 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.496363 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.526684 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.536726 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.590486 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.612466 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.694619 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.724679 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.752029 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.779321 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.809755 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.841007 4744 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.927722 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.970803 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:22:27 crc kubenswrapper[4744]: I1201 08:22:27.990917 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.024367 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.064771 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.066047 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.093102 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.104768 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.319223 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.432934 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.441198 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.483206 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.540649 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.609202 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.626610 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.717617 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.807016 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.872663 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.880431 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.931531 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.947215 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 08:22:28 crc kubenswrapper[4744]: I1201 08:22:28.998500 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.035857 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.156477 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.259181 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.262146 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.289647 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.300992 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.371803 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.394366 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.402613 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.439090 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.503845 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.857525 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.857716 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.858025 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.858281 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.858424 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.858538 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.860046 4744 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.872568 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.880472 4744 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.884322 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-pflgv"] Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.884378 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.891557 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.903790 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.90377358 podStartE2EDuration="17.90377358s" podCreationTimestamp="2025-12-01 08:22:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:22:29.899341238 +0000 UTC m=+281.888399169" watchObservedRunningTime="2025-12-01 08:22:29.90377358 +0000 UTC m=+281.892831491" Dec 01 08:22:29 crc kubenswrapper[4744]: I1201 08:22:29.990227 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.097609 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.159129 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.179696 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.296188 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.296713 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" path="/var/lib/kubelet/pods/34c76ecb-9f6d-4c8e-b147-5ac1b981cb61/volumes" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.332843 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.347661 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.449602 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.502954 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.654231 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 08:22:30 crc kubenswrapper[4744]: I1201 08:22:30.729766 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.042678 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.091016 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.171106 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.200233 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.312166 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.317346 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.318651 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.333523 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.381928 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.403061 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.483963 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.508905 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.738473 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.793289 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.823757 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.843908 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.863272 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.908222 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 08:22:31 crc kubenswrapper[4744]: I1201 08:22:31.963928 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.036228 4744 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.044135 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.067781 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.075788 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.154866 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.157854 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.199849 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.235367 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.344115 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.430898 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.443877 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.469312 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.482274 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.483129 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.503279 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.515182 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.562400 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.768885 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.926680 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 08:22:32 crc kubenswrapper[4744]: I1201 08:22:32.953111 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.038820 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.122239 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.253877 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.328248 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.346274 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.359078 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.403701 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.440530 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.474879 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.542305 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.555563 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.624892 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.728592 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.764846 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.851737 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.900227 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.951491 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:22:33 crc kubenswrapper[4744]: I1201 08:22:33.992769 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.009907 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.083625 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.414592 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.493193 4744 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.493434 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09" gracePeriod=5 Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.540914 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.608714 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.609709 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.652438 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.729668 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.749664 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.787805 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.865692 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.898521 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.915465 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 08:22:34 crc kubenswrapper[4744]: I1201 08:22:34.927320 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.028719 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.037230 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.067429 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.068814 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.150280 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.225185 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.280285 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.281599 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.409870 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.417813 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.472929 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.484758 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.495673 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.568548 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.825990 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.853843 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.932392 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 08:22:35 crc kubenswrapper[4744]: I1201 08:22:35.955856 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.153077 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.156551 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.251266 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.319558 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.322926 4744 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.484539 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.526318 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.624073 4744 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.796624 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.838923 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 08:22:36 crc kubenswrapper[4744]: I1201 08:22:36.886634 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 08:22:37 crc kubenswrapper[4744]: I1201 08:22:37.012305 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:22:37 crc kubenswrapper[4744]: I1201 08:22:37.294045 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 08:22:37 crc kubenswrapper[4744]: I1201 08:22:37.308044 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 08:22:37 crc kubenswrapper[4744]: I1201 08:22:37.494790 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 08:22:37 crc kubenswrapper[4744]: I1201 08:22:37.919752 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.169016 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.221386 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.228682 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.336819 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.410630 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.459634 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.558870 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.603775 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.645321 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.833708 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.856301 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 08:22:38 crc kubenswrapper[4744]: I1201 08:22:38.914918 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 08:22:39 crc kubenswrapper[4744]: I1201 08:22:39.071885 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 08:22:39 crc kubenswrapper[4744]: I1201 08:22:39.189263 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 08:22:39 crc kubenswrapper[4744]: I1201 08:22:39.513297 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.090640 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.091537 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.265082 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.265792 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.266055 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.265970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.266283 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.266446 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.266293 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.266988 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.267273 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.267851 4744 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.268079 4744 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.268275 4744 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.268494 4744 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.281868 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.295702 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.337984 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.338084 4744 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09" exitCode=137 Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.338159 4744 scope.go:117] "RemoveContainer" containerID="674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.338194 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.363244 4744 scope.go:117] "RemoveContainer" containerID="674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09" Dec 01 08:22:40 crc kubenswrapper[4744]: E1201 08:22:40.364126 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09\": container with ID starting with 674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09 not found: ID does not exist" containerID="674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.364181 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09"} err="failed to get container status \"674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09\": rpc error: code = NotFound desc = could not find container \"674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09\": container with ID starting with 674db08513eaf2aa95608267fe011efb0557645b7c4da7c658647950c4e4ca09 not found: ID does not exist" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.370300 4744 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.576460 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-79558c6cc6-4m5lv"] Dec 01 08:22:40 crc kubenswrapper[4744]: E1201 08:22:40.576779 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" containerName="oauth-openshift" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.576801 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" containerName="oauth-openshift" Dec 01 08:22:40 crc kubenswrapper[4744]: E1201 08:22:40.576830 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" containerName="installer" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.576843 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" containerName="installer" Dec 01 08:22:40 crc kubenswrapper[4744]: E1201 08:22:40.576863 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.576876 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.577037 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e227ac63-afad-4e84-937d-98f39aa3d985" containerName="installer" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.577062 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.577082 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c76ecb-9f6d-4c8e-b147-5ac1b981cb61" containerName="oauth-openshift" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.577696 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.581100 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.581582 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.582323 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.582977 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.584731 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.585648 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.585727 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.586026 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.586049 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.586450 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.586561 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.587525 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.600817 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.608645 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.613947 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.618976 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79558c6cc6-4m5lv"] Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.674888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-service-ca\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.674988 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675040 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7nhf\" (UniqueName: \"kubernetes.io/projected/89c2c301-5470-498b-8ad6-399b57597f57-kube-api-access-c7nhf\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675077 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-audit-policies\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675125 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-session\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675185 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675231 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-login\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675371 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675477 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-router-certs\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675513 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-error\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675553 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89c2c301-5470-498b-8ad6-399b57597f57-audit-dir\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675607 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.675652 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.776926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-service-ca\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777023 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7nhf\" (UniqueName: \"kubernetes.io/projected/89c2c301-5470-498b-8ad6-399b57597f57-kube-api-access-c7nhf\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-audit-policies\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-session\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777223 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-login\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777388 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777471 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-router-certs\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777555 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-error\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777602 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89c2c301-5470-498b-8ad6-399b57597f57-audit-dir\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777655 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.777703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.778349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89c2c301-5470-498b-8ad6-399b57597f57-audit-dir\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.778700 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-audit-policies\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.778759 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-service-ca\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.778921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.779704 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.784970 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-error\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.785294 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.785369 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-router-certs\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.785490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-session\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.786350 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-login\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.787603 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.796156 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.797038 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89c2c301-5470-498b-8ad6-399b57597f57-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.811134 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7nhf\" (UniqueName: \"kubernetes.io/projected/89c2c301-5470-498b-8ad6-399b57597f57-kube-api-access-c7nhf\") pod \"oauth-openshift-79558c6cc6-4m5lv\" (UID: \"89c2c301-5470-498b-8ad6-399b57597f57\") " pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:40 crc kubenswrapper[4744]: I1201 08:22:40.906595 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:41 crc kubenswrapper[4744]: I1201 08:22:41.238025 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79558c6cc6-4m5lv"] Dec 01 08:22:41 crc kubenswrapper[4744]: I1201 08:22:41.353912 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" event={"ID":"89c2c301-5470-498b-8ad6-399b57597f57","Type":"ContainerStarted","Data":"d9402aa7698da2aa8346e1083b43b206ebf2e62a6f8003455177b529d0555f7a"} Dec 01 08:22:42 crc kubenswrapper[4744]: I1201 08:22:42.365445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" event={"ID":"89c2c301-5470-498b-8ad6-399b57597f57","Type":"ContainerStarted","Data":"1525ffaa237b1a064d7978060f142a8241b5a0dbadb4a32f52354a3b423881ea"} Dec 01 08:22:42 crc kubenswrapper[4744]: I1201 08:22:42.365834 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:42 crc kubenswrapper[4744]: I1201 08:22:42.380933 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" Dec 01 08:22:42 crc kubenswrapper[4744]: I1201 08:22:42.401758 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-79558c6cc6-4m5lv" podStartSLOduration=68.401731737 podStartE2EDuration="1m8.401731737s" podCreationTimestamp="2025-12-01 08:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:22:42.401462009 +0000 UTC m=+294.390519970" watchObservedRunningTime="2025-12-01 08:22:42.401731737 +0000 UTC m=+294.390789688" Dec 01 08:22:48 crc kubenswrapper[4744]: I1201 08:22:48.161883 4744 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.114455 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pnzr5"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.115522 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pnzr5" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="registry-server" containerID="cri-o://aad146934f9e5a08e12f5151fa75f724b41ac37c0e58dde86b827407ade64f2c" gracePeriod=30 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.124329 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pvtt2"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.126060 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pvtt2" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="registry-server" containerID="cri-o://b8ef7f269fd4b2371a93ad9795f22e8d16bf21806a0b2642c4d534003700caea" gracePeriod=30 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.135112 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qpjtj"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.142837 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" podUID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" containerName="marketplace-operator" containerID="cri-o://2787b08f1c7d6efadb919e2b8925c3377a2bb9de24214ca1e3cf5962992643f2" gracePeriod=30 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.145121 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rscr"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.145338 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6rscr" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="registry-server" containerID="cri-o://48d15cab9941a94430f5fae348bbecd90592211f8107b9956c65be7e206cbbf3" gracePeriod=30 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.173132 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gs9pb"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.173537 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gs9pb" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="registry-server" containerID="cri-o://6325a8ee70f509963ff6c9d9dcadcbe620b52d357e95b1ce0229e2e4cae1aa45" gracePeriod=30 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.188747 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-99jdl"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.189899 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.206788 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-99jdl"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.320674 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqhqf\" (UniqueName: \"kubernetes.io/projected/94e841b2-c9ed-43c3-b96c-07661415a958-kube-api-access-nqhqf\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.321005 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94e841b2-c9ed-43c3-b96c-07661415a958-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.321060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e841b2-c9ed-43c3-b96c-07661415a958-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.422956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqhqf\" (UniqueName: \"kubernetes.io/projected/94e841b2-c9ed-43c3-b96c-07661415a958-kube-api-access-nqhqf\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.423816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94e841b2-c9ed-43c3-b96c-07661415a958-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.424613 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e841b2-c9ed-43c3-b96c-07661415a958-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.426168 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e841b2-c9ed-43c3-b96c-07661415a958-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.426216 4744 generic.go:334] "Generic (PLEG): container finished" podID="fe4049b3-153a-45d1-8847-a51f91053f77" containerID="6325a8ee70f509963ff6c9d9dcadcbe620b52d357e95b1ce0229e2e4cae1aa45" exitCode=0 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.426305 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs9pb" event={"ID":"fe4049b3-153a-45d1-8847-a51f91053f77","Type":"ContainerDied","Data":"6325a8ee70f509963ff6c9d9dcadcbe620b52d357e95b1ce0229e2e4cae1aa45"} Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.427867 4744 generic.go:334] "Generic (PLEG): container finished" podID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" containerID="2787b08f1c7d6efadb919e2b8925c3377a2bb9de24214ca1e3cf5962992643f2" exitCode=0 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.427955 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" event={"ID":"948afa07-87fa-4c8d-a26d-5b9b8d3b156a","Type":"ContainerDied","Data":"2787b08f1c7d6efadb919e2b8925c3377a2bb9de24214ca1e3cf5962992643f2"} Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.430057 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94e841b2-c9ed-43c3-b96c-07661415a958-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.435870 4744 generic.go:334] "Generic (PLEG): container finished" podID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerID="aad146934f9e5a08e12f5151fa75f724b41ac37c0e58dde86b827407ade64f2c" exitCode=0 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.435919 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnzr5" event={"ID":"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b","Type":"ContainerDied","Data":"aad146934f9e5a08e12f5151fa75f724b41ac37c0e58dde86b827407ade64f2c"} Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.439049 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqhqf\" (UniqueName: \"kubernetes.io/projected/94e841b2-c9ed-43c3-b96c-07661415a958-kube-api-access-nqhqf\") pod \"marketplace-operator-79b997595-99jdl\" (UID: \"94e841b2-c9ed-43c3-b96c-07661415a958\") " pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.440688 4744 generic.go:334] "Generic (PLEG): container finished" podID="862392cd-33ed-4e07-8ea5-eef33e192765" containerID="48d15cab9941a94430f5fae348bbecd90592211f8107b9956c65be7e206cbbf3" exitCode=0 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.440768 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rscr" event={"ID":"862392cd-33ed-4e07-8ea5-eef33e192765","Type":"ContainerDied","Data":"48d15cab9941a94430f5fae348bbecd90592211f8107b9956c65be7e206cbbf3"} Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.443176 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerID="b8ef7f269fd4b2371a93ad9795f22e8d16bf21806a0b2642c4d534003700caea" exitCode=0 Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.443221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvtt2" event={"ID":"d2e302f9-45f6-4e8e-9c08-984faf8732d5","Type":"ContainerDied","Data":"b8ef7f269fd4b2371a93ad9795f22e8d16bf21806a0b2642c4d534003700caea"} Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.604795 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.610608 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.619265 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.638567 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.668108 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.699159 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729002 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-utilities\") pod \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729040 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-catalog-content\") pod \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729084 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-kube-api-access-5n6zx\") pod \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729105 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-operator-metrics\") pod \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729160 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdklm\" (UniqueName: \"kubernetes.io/projected/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-kube-api-access-rdklm\") pod \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\" (UID: \"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729194 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-utilities\") pod \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729215 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-catalog-content\") pod \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729235 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msgnj\" (UniqueName: \"kubernetes.io/projected/d2e302f9-45f6-4e8e-9c08-984faf8732d5-kube-api-access-msgnj\") pod \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\" (UID: \"d2e302f9-45f6-4e8e-9c08-984faf8732d5\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.729255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-trusted-ca\") pod \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\" (UID: \"948afa07-87fa-4c8d-a26d-5b9b8d3b156a\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.730633 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-utilities" (OuterVolumeSpecName: "utilities") pod "d2e302f9-45f6-4e8e-9c08-984faf8732d5" (UID: "d2e302f9-45f6-4e8e-9c08-984faf8732d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.730685 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "948afa07-87fa-4c8d-a26d-5b9b8d3b156a" (UID: "948afa07-87fa-4c8d-a26d-5b9b8d3b156a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.731463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-utilities" (OuterVolumeSpecName: "utilities") pod "6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" (UID: "6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.733922 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e302f9-45f6-4e8e-9c08-984faf8732d5-kube-api-access-msgnj" (OuterVolumeSpecName: "kube-api-access-msgnj") pod "d2e302f9-45f6-4e8e-9c08-984faf8732d5" (UID: "d2e302f9-45f6-4e8e-9c08-984faf8732d5"). InnerVolumeSpecName "kube-api-access-msgnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.734896 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-kube-api-access-5n6zx" (OuterVolumeSpecName: "kube-api-access-5n6zx") pod "948afa07-87fa-4c8d-a26d-5b9b8d3b156a" (UID: "948afa07-87fa-4c8d-a26d-5b9b8d3b156a"). InnerVolumeSpecName "kube-api-access-5n6zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.734927 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "948afa07-87fa-4c8d-a26d-5b9b8d3b156a" (UID: "948afa07-87fa-4c8d-a26d-5b9b8d3b156a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.735573 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-kube-api-access-rdklm" (OuterVolumeSpecName: "kube-api-access-rdklm") pod "6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" (UID: "6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b"). InnerVolumeSpecName "kube-api-access-rdklm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.793434 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" (UID: "6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.815001 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2e302f9-45f6-4e8e-9c08-984faf8732d5" (UID: "d2e302f9-45f6-4e8e-9c08-984faf8732d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.829918 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5xgf\" (UniqueName: \"kubernetes.io/projected/fe4049b3-153a-45d1-8847-a51f91053f77-kube-api-access-z5xgf\") pod \"fe4049b3-153a-45d1-8847-a51f91053f77\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.829983 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-utilities\") pod \"fe4049b3-153a-45d1-8847-a51f91053f77\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxd8c\" (UniqueName: \"kubernetes.io/projected/862392cd-33ed-4e07-8ea5-eef33e192765-kube-api-access-dxd8c\") pod \"862392cd-33ed-4e07-8ea5-eef33e192765\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-catalog-content\") pod \"fe4049b3-153a-45d1-8847-a51f91053f77\" (UID: \"fe4049b3-153a-45d1-8847-a51f91053f77\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830085 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-catalog-content\") pod \"862392cd-33ed-4e07-8ea5-eef33e192765\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830124 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-utilities\") pod \"862392cd-33ed-4e07-8ea5-eef33e192765\" (UID: \"862392cd-33ed-4e07-8ea5-eef33e192765\") " Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830355 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830368 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830377 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830385 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-kube-api-access-5n6zx\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830393 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/948afa07-87fa-4c8d-a26d-5b9b8d3b156a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830401 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdklm\" (UniqueName: \"kubernetes.io/projected/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b-kube-api-access-rdklm\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830427 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830434 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e302f9-45f6-4e8e-9c08-984faf8732d5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.830443 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msgnj\" (UniqueName: \"kubernetes.io/projected/d2e302f9-45f6-4e8e-9c08-984faf8732d5-kube-api-access-msgnj\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.831060 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-utilities" (OuterVolumeSpecName: "utilities") pod "862392cd-33ed-4e07-8ea5-eef33e192765" (UID: "862392cd-33ed-4e07-8ea5-eef33e192765"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.831202 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-utilities" (OuterVolumeSpecName: "utilities") pod "fe4049b3-153a-45d1-8847-a51f91053f77" (UID: "fe4049b3-153a-45d1-8847-a51f91053f77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.832948 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe4049b3-153a-45d1-8847-a51f91053f77-kube-api-access-z5xgf" (OuterVolumeSpecName: "kube-api-access-z5xgf") pod "fe4049b3-153a-45d1-8847-a51f91053f77" (UID: "fe4049b3-153a-45d1-8847-a51f91053f77"). InnerVolumeSpecName "kube-api-access-z5xgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.833133 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-99jdl"] Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.833570 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862392cd-33ed-4e07-8ea5-eef33e192765-kube-api-access-dxd8c" (OuterVolumeSpecName: "kube-api-access-dxd8c") pod "862392cd-33ed-4e07-8ea5-eef33e192765" (UID: "862392cd-33ed-4e07-8ea5-eef33e192765"). InnerVolumeSpecName "kube-api-access-dxd8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.847159 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "862392cd-33ed-4e07-8ea5-eef33e192765" (UID: "862392cd-33ed-4e07-8ea5-eef33e192765"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.931381 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5xgf\" (UniqueName: \"kubernetes.io/projected/fe4049b3-153a-45d1-8847-a51f91053f77-kube-api-access-z5xgf\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.931448 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.931460 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxd8c\" (UniqueName: \"kubernetes.io/projected/862392cd-33ed-4e07-8ea5-eef33e192765-kube-api-access-dxd8c\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.931471 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.931481 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/862392cd-33ed-4e07-8ea5-eef33e192765-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:50 crc kubenswrapper[4744]: I1201 08:22:50.933426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe4049b3-153a-45d1-8847-a51f91053f77" (UID: "fe4049b3-153a-45d1-8847-a51f91053f77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.032275 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4049b3-153a-45d1-8847-a51f91053f77-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.466923 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnzr5" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.467284 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnzr5" event={"ID":"6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b","Type":"ContainerDied","Data":"5045d5b2f6159e439f505fa8b1bd8c65cbd671a3807f58e103c3afcd64f6d0fe"} Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.467401 4744 scope.go:117] "RemoveContainer" containerID="aad146934f9e5a08e12f5151fa75f724b41ac37c0e58dde86b827407ade64f2c" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.472739 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rscr" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.473077 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rscr" event={"ID":"862392cd-33ed-4e07-8ea5-eef33e192765","Type":"ContainerDied","Data":"e24adbee2855d9ca517a1ce836111b048190cdb9366826927d2fa54e5285bf54"} Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.475528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" event={"ID":"94e841b2-c9ed-43c3-b96c-07661415a958","Type":"ContainerStarted","Data":"d6cd6f1e8b7ed74b70df4d558dc37c9ea5a30c9817eeb6b4860557f0307ebaf2"} Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.475599 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" event={"ID":"94e841b2-c9ed-43c3-b96c-07661415a958","Type":"ContainerStarted","Data":"83fa70863b15fcdffb9669896ccf5056d7c96d60377b01359dca1f696a369f70"} Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.475745 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.478374 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvtt2" event={"ID":"d2e302f9-45f6-4e8e-9c08-984faf8732d5","Type":"ContainerDied","Data":"04bcc6af421a415dcfb1e7dc3df554233069eeab318ffccad577d330678eac99"} Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.478528 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvtt2" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.480049 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.483227 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs9pb" event={"ID":"fe4049b3-153a-45d1-8847-a51f91053f77","Type":"ContainerDied","Data":"bb41122cb3450bdf6cafcf9935f7417deb02d176125e057767480ea83d05d6ea"} Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.483388 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs9pb" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.487028 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" event={"ID":"948afa07-87fa-4c8d-a26d-5b9b8d3b156a","Type":"ContainerDied","Data":"de8ebdd6b6c3c4b8aa46fafa2efde83ffb9b981e91b0b075aa62ca997016b8b3"} Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.487104 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qpjtj" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.499376 4744 scope.go:117] "RemoveContainer" containerID="c4248cbb3325cec506a610f3e7dba2afc7fb509400f322ab13a5395435c6efdf" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.506655 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-99jdl" podStartSLOduration=1.506618955 podStartE2EDuration="1.506618955s" podCreationTimestamp="2025-12-01 08:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:22:51.498017029 +0000 UTC m=+303.487074960" watchObservedRunningTime="2025-12-01 08:22:51.506618955 +0000 UTC m=+303.495676916" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.555274 4744 scope.go:117] "RemoveContainer" containerID="63bb2f481bdf53d281e5dea3a0480333e24c406e8a2ae53ee2f2b2cc08dc00c2" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.565045 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pvtt2"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.574531 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pvtt2"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.577275 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pnzr5"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.583341 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pnzr5"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.599670 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rscr"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.604201 4744 scope.go:117] "RemoveContainer" containerID="48d15cab9941a94430f5fae348bbecd90592211f8107b9956c65be7e206cbbf3" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.606257 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rscr"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.611165 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qpjtj"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.615133 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qpjtj"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.624108 4744 scope.go:117] "RemoveContainer" containerID="c388bb3caf9de93279d4e066ba6b9a930f3321000920d2b96f1f37e7904ca741" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.630217 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gs9pb"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.633461 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gs9pb"] Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.647349 4744 scope.go:117] "RemoveContainer" containerID="7155be193e36537589e6d80d1e8fb08b398ef3f23e04d6b349147069af172b83" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.664034 4744 scope.go:117] "RemoveContainer" containerID="b8ef7f269fd4b2371a93ad9795f22e8d16bf21806a0b2642c4d534003700caea" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.691078 4744 scope.go:117] "RemoveContainer" containerID="85fc8e77989cd48773ebb98c7fb77218c0ea6e79718443322dd998438d034b06" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.715352 4744 scope.go:117] "RemoveContainer" containerID="f8955fd17e87e863e87f3143141d3542f40c2f19348f5913c14e3ddd3a6ba04f" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.734216 4744 scope.go:117] "RemoveContainer" containerID="6325a8ee70f509963ff6c9d9dcadcbe620b52d357e95b1ce0229e2e4cae1aa45" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.750306 4744 scope.go:117] "RemoveContainer" containerID="d0c9a6e85783b095e2a3a26386797dd4499efa2bcd36990b19b3ebda43bda198" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.768781 4744 scope.go:117] "RemoveContainer" containerID="b591cfe080e5743432c243050dbb66b23888d0ec6a9fe0c87ae05755d1b1e0da" Dec 01 08:22:51 crc kubenswrapper[4744]: I1201 08:22:51.782551 4744 scope.go:117] "RemoveContainer" containerID="2787b08f1c7d6efadb919e2b8925c3377a2bb9de24214ca1e3cf5962992643f2" Dec 01 08:22:52 crc kubenswrapper[4744]: I1201 08:22:52.291503 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" path="/var/lib/kubelet/pods/6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b/volumes" Dec 01 08:22:52 crc kubenswrapper[4744]: I1201 08:22:52.292229 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" path="/var/lib/kubelet/pods/862392cd-33ed-4e07-8ea5-eef33e192765/volumes" Dec 01 08:22:52 crc kubenswrapper[4744]: I1201 08:22:52.292813 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" path="/var/lib/kubelet/pods/948afa07-87fa-4c8d-a26d-5b9b8d3b156a/volumes" Dec 01 08:22:52 crc kubenswrapper[4744]: I1201 08:22:52.293252 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" path="/var/lib/kubelet/pods/d2e302f9-45f6-4e8e-9c08-984faf8732d5/volumes" Dec 01 08:22:52 crc kubenswrapper[4744]: I1201 08:22:52.293807 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" path="/var/lib/kubelet/pods/fe4049b3-153a-45d1-8847-a51f91053f77/volumes" Dec 01 08:23:07 crc kubenswrapper[4744]: I1201 08:23:07.667563 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6qsl"] Dec 01 08:23:07 crc kubenswrapper[4744]: I1201 08:23:07.668237 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" podUID="a87bf7b3-7b9d-407d-a113-387afac58cab" containerName="controller-manager" containerID="cri-o://c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de" gracePeriod=30 Dec 01 08:23:07 crc kubenswrapper[4744]: I1201 08:23:07.774144 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s"] Dec 01 08:23:07 crc kubenswrapper[4744]: I1201 08:23:07.774391 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" podUID="268ee28c-0643-4161-aee8-6fa5a08722c8" containerName="route-controller-manager" containerID="cri-o://c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00" gracePeriod=30 Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.095898 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.181190 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.272962 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-config\") pod \"268ee28c-0643-4161-aee8-6fa5a08722c8\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.281965 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/268ee28c-0643-4161-aee8-6fa5a08722c8-serving-cert\") pod \"268ee28c-0643-4161-aee8-6fa5a08722c8\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.282011 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87bf7b3-7b9d-407d-a113-387afac58cab-serving-cert\") pod \"a87bf7b3-7b9d-407d-a113-387afac58cab\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.282035 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-client-ca\") pod \"268ee28c-0643-4161-aee8-6fa5a08722c8\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.282108 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-config\") pod \"a87bf7b3-7b9d-407d-a113-387afac58cab\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.282141 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-client-ca\") pod \"a87bf7b3-7b9d-407d-a113-387afac58cab\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.282159 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4bb8\" (UniqueName: \"kubernetes.io/projected/268ee28c-0643-4161-aee8-6fa5a08722c8-kube-api-access-q4bb8\") pod \"268ee28c-0643-4161-aee8-6fa5a08722c8\" (UID: \"268ee28c-0643-4161-aee8-6fa5a08722c8\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.282184 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-proxy-ca-bundles\") pod \"a87bf7b3-7b9d-407d-a113-387afac58cab\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.282203 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd87x\" (UniqueName: \"kubernetes.io/projected/a87bf7b3-7b9d-407d-a113-387afac58cab-kube-api-access-rd87x\") pod \"a87bf7b3-7b9d-407d-a113-387afac58cab\" (UID: \"a87bf7b3-7b9d-407d-a113-387afac58cab\") " Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.279810 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-config" (OuterVolumeSpecName: "config") pod "268ee28c-0643-4161-aee8-6fa5a08722c8" (UID: "268ee28c-0643-4161-aee8-6fa5a08722c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.284722 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-client-ca" (OuterVolumeSpecName: "client-ca") pod "268ee28c-0643-4161-aee8-6fa5a08722c8" (UID: "268ee28c-0643-4161-aee8-6fa5a08722c8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.287458 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a87bf7b3-7b9d-407d-a113-387afac58cab" (UID: "a87bf7b3-7b9d-407d-a113-387afac58cab"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.287571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-client-ca" (OuterVolumeSpecName: "client-ca") pod "a87bf7b3-7b9d-407d-a113-387afac58cab" (UID: "a87bf7b3-7b9d-407d-a113-387afac58cab"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.287601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-config" (OuterVolumeSpecName: "config") pod "a87bf7b3-7b9d-407d-a113-387afac58cab" (UID: "a87bf7b3-7b9d-407d-a113-387afac58cab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.290217 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268ee28c-0643-4161-aee8-6fa5a08722c8-kube-api-access-q4bb8" (OuterVolumeSpecName: "kube-api-access-q4bb8") pod "268ee28c-0643-4161-aee8-6fa5a08722c8" (UID: "268ee28c-0643-4161-aee8-6fa5a08722c8"). InnerVolumeSpecName "kube-api-access-q4bb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.290371 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268ee28c-0643-4161-aee8-6fa5a08722c8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "268ee28c-0643-4161-aee8-6fa5a08722c8" (UID: "268ee28c-0643-4161-aee8-6fa5a08722c8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.290450 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87bf7b3-7b9d-407d-a113-387afac58cab-kube-api-access-rd87x" (OuterVolumeSpecName: "kube-api-access-rd87x") pod "a87bf7b3-7b9d-407d-a113-387afac58cab" (UID: "a87bf7b3-7b9d-407d-a113-387afac58cab"). InnerVolumeSpecName "kube-api-access-rd87x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.290878 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87bf7b3-7b9d-407d-a113-387afac58cab-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a87bf7b3-7b9d-407d-a113-387afac58cab" (UID: "a87bf7b3-7b9d-407d-a113-387afac58cab"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384284 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384340 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4bb8\" (UniqueName: \"kubernetes.io/projected/268ee28c-0643-4161-aee8-6fa5a08722c8-kube-api-access-q4bb8\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384361 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384380 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd87x\" (UniqueName: \"kubernetes.io/projected/a87bf7b3-7b9d-407d-a113-387afac58cab-kube-api-access-rd87x\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384398 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384439 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/268ee28c-0643-4161-aee8-6fa5a08722c8-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384456 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87bf7b3-7b9d-407d-a113-387afac58cab-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384472 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/268ee28c-0643-4161-aee8-6fa5a08722c8-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.384488 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87bf7b3-7b9d-407d-a113-387afac58cab-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.611144 4744 generic.go:334] "Generic (PLEG): container finished" podID="268ee28c-0643-4161-aee8-6fa5a08722c8" containerID="c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00" exitCode=0 Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.611205 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" event={"ID":"268ee28c-0643-4161-aee8-6fa5a08722c8","Type":"ContainerDied","Data":"c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00"} Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.611262 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" event={"ID":"268ee28c-0643-4161-aee8-6fa5a08722c8","Type":"ContainerDied","Data":"df7f6e0302cc19d9f5ec354e76010432c767cdf876af63d755d3e873336a1b27"} Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.611262 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.611289 4744 scope.go:117] "RemoveContainer" containerID="c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.614903 4744 generic.go:334] "Generic (PLEG): container finished" podID="a87bf7b3-7b9d-407d-a113-387afac58cab" containerID="c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de" exitCode=0 Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.614935 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" event={"ID":"a87bf7b3-7b9d-407d-a113-387afac58cab","Type":"ContainerDied","Data":"c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de"} Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.614958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" event={"ID":"a87bf7b3-7b9d-407d-a113-387afac58cab","Type":"ContainerDied","Data":"fc2d0e363778c0ee556e8c3b9d1ffafb4781b225f5ac4f139ee908064342096b"} Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.615011 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r6qsl" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.638528 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6qsl"] Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.643875 4744 scope.go:117] "RemoveContainer" containerID="c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.644338 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r6qsl"] Dec 01 08:23:08 crc kubenswrapper[4744]: E1201 08:23:08.644391 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00\": container with ID starting with c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00 not found: ID does not exist" containerID="c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.644458 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00"} err="failed to get container status \"c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00\": rpc error: code = NotFound desc = could not find container \"c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00\": container with ID starting with c8e3c4f8bb894b17ee07d4555d8ddfbeeedee67ea93cf742a796f6867ca87f00 not found: ID does not exist" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.644486 4744 scope.go:117] "RemoveContainer" containerID="c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.653168 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s"] Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.658000 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-nr58s"] Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.664682 4744 scope.go:117] "RemoveContainer" containerID="c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de" Dec 01 08:23:08 crc kubenswrapper[4744]: E1201 08:23:08.665328 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de\": container with ID starting with c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de not found: ID does not exist" containerID="c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de" Dec 01 08:23:08 crc kubenswrapper[4744]: I1201 08:23:08.665388 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de"} err="failed to get container status \"c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de\": rpc error: code = NotFound desc = could not find container \"c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de\": container with ID starting with c149004606243fffdcaa57533e51c31936492ffb301af9dc6ee1393c11a2a8de not found: ID does not exist" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597266 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9"] Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.597782 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597806 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.597832 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597844 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.597861 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597876 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.597891 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268ee28c-0643-4161-aee8-6fa5a08722c8" containerName="route-controller-manager" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597903 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="268ee28c-0643-4161-aee8-6fa5a08722c8" containerName="route-controller-manager" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.597918 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597930 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.597943 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597956 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.597976 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.597987 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598007 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598020 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598034 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598046 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598060 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598072 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598088 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598099 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="extract-content" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598118 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" containerName="marketplace-operator" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598130 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" containerName="marketplace-operator" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598147 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598161 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="extract-utilities" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598175 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87bf7b3-7b9d-407d-a113-387afac58cab" containerName="controller-manager" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598187 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87bf7b3-7b9d-407d-a113-387afac58cab" containerName="controller-manager" Dec 01 08:23:09 crc kubenswrapper[4744]: E1201 08:23:09.598202 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598214 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598383 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="268ee28c-0643-4161-aee8-6fa5a08722c8" containerName="route-controller-manager" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598402 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e302f9-45f6-4e8e-9c08-984faf8732d5" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598439 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe4049b3-153a-45d1-8847-a51f91053f77" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598452 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="948afa07-87fa-4c8d-a26d-5b9b8d3b156a" containerName="marketplace-operator" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598469 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fdf6ae8-2821-49f9-9ecf-9c4b77d5a56b" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598484 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87bf7b3-7b9d-407d-a113-387afac58cab" containerName="controller-manager" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.598502 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="862392cd-33ed-4e07-8ea5-eef33e192765" containerName="registry-server" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.601582 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.604813 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.605187 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.605680 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.606010 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.610174 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f9cf784b-b4xxq"] Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.613981 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.616564 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9"] Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.628954 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.629312 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.632054 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.632319 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.632687 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.633072 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.633871 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.634236 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.638323 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.683486 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f9cf784b-b4xxq"] Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703190 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-proxy-ca-bundles\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703303 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-config\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703335 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7n5p\" (UniqueName: \"kubernetes.io/projected/e7e864a6-456d-43e4-a6fa-f774f36c658b-kube-api-access-m7n5p\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-client-ca\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fc08073-673a-4dc6-a70e-20f850dbb988-serving-cert\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-config\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703496 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e864a6-456d-43e4-a6fa-f774f36c658b-serving-cert\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khm88\" (UniqueName: \"kubernetes.io/projected/2fc08073-673a-4dc6-a70e-20f850dbb988-kube-api-access-khm88\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.703589 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-client-ca\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805332 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-config\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805402 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7n5p\" (UniqueName: \"kubernetes.io/projected/e7e864a6-456d-43e4-a6fa-f774f36c658b-kube-api-access-m7n5p\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-client-ca\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805579 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fc08073-673a-4dc6-a70e-20f850dbb988-serving-cert\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-config\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805699 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e864a6-456d-43e4-a6fa-f774f36c658b-serving-cert\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805789 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khm88\" (UniqueName: \"kubernetes.io/projected/2fc08073-673a-4dc6-a70e-20f850dbb988-kube-api-access-khm88\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-client-ca\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.805923 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-proxy-ca-bundles\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.806463 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-client-ca\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.807590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-client-ca\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.807891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-config\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.809148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-config\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.809309 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7e864a6-456d-43e4-a6fa-f774f36c658b-proxy-ca-bundles\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.809973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fc08073-673a-4dc6-a70e-20f850dbb988-serving-cert\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.810711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e864a6-456d-43e4-a6fa-f774f36c658b-serving-cert\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.830556 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7n5p\" (UniqueName: \"kubernetes.io/projected/e7e864a6-456d-43e4-a6fa-f774f36c658b-kube-api-access-m7n5p\") pod \"controller-manager-f9cf784b-b4xxq\" (UID: \"e7e864a6-456d-43e4-a6fa-f774f36c658b\") " pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.834184 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khm88\" (UniqueName: \"kubernetes.io/projected/2fc08073-673a-4dc6-a70e-20f850dbb988-kube-api-access-khm88\") pod \"route-controller-manager-78d46db78f-zt9b9\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.978061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:09 crc kubenswrapper[4744]: I1201 08:23:09.986480 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.201887 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f9cf784b-b4xxq"] Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.231460 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9"] Dec 01 08:23:10 crc kubenswrapper[4744]: W1201 08:23:10.245421 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fc08073_673a_4dc6_a70e_20f850dbb988.slice/crio-de9629e5f98640469a00587e513e056ef2e180ca7e0c96e1a4d9f28df6ed586d WatchSource:0}: Error finding container de9629e5f98640469a00587e513e056ef2e180ca7e0c96e1a4d9f28df6ed586d: Status 404 returned error can't find the container with id de9629e5f98640469a00587e513e056ef2e180ca7e0c96e1a4d9f28df6ed586d Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.298892 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="268ee28c-0643-4161-aee8-6fa5a08722c8" path="/var/lib/kubelet/pods/268ee28c-0643-4161-aee8-6fa5a08722c8/volumes" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.300152 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a87bf7b3-7b9d-407d-a113-387afac58cab" path="/var/lib/kubelet/pods/a87bf7b3-7b9d-407d-a113-387afac58cab/volumes" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.660385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" event={"ID":"e7e864a6-456d-43e4-a6fa-f774f36c658b","Type":"ContainerStarted","Data":"35a1e4922e7aed79f2caa935ae14ec91f5dadce84963860bcc2b3a283ab27bb7"} Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.660441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" event={"ID":"e7e864a6-456d-43e4-a6fa-f774f36c658b","Type":"ContainerStarted","Data":"f8414cb0ef173c18598d20704841da1d211ecf5c98a9394024cef9e6a2376bc3"} Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.660578 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.661852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" event={"ID":"2fc08073-673a-4dc6-a70e-20f850dbb988","Type":"ContainerStarted","Data":"882618f84cf5a287981fbf6984d8baed0720154c2aef509511915689166d78f6"} Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.661875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" event={"ID":"2fc08073-673a-4dc6-a70e-20f850dbb988","Type":"ContainerStarted","Data":"de9629e5f98640469a00587e513e056ef2e180ca7e0c96e1a4d9f28df6ed586d"} Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.662495 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.666045 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.673330 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.706603 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f9cf784b-b4xxq" podStartSLOduration=3.706585575 podStartE2EDuration="3.706585575s" podCreationTimestamp="2025-12-01 08:23:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:23:10.681547402 +0000 UTC m=+322.670605323" watchObservedRunningTime="2025-12-01 08:23:10.706585575 +0000 UTC m=+322.695643496" Dec 01 08:23:10 crc kubenswrapper[4744]: I1201 08:23:10.767741 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" podStartSLOduration=3.767725375 podStartE2EDuration="3.767725375s" podCreationTimestamp="2025-12-01 08:23:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:23:10.765950058 +0000 UTC m=+322.755007969" watchObservedRunningTime="2025-12-01 08:23:10.767725375 +0000 UTC m=+322.756783296" Dec 01 08:23:12 crc kubenswrapper[4744]: I1201 08:23:12.998597 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wd2vs"] Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.001159 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.005158 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.019632 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wd2vs"] Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.098670 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8744\" (UniqueName: \"kubernetes.io/projected/a5519f09-e6ba-4df8-8d28-d014f65fff6f-kube-api-access-b8744\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.098756 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5519f09-e6ba-4df8-8d28-d014f65fff6f-utilities\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.098943 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5519f09-e6ba-4df8-8d28-d014f65fff6f-catalog-content\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.184261 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5lhd"] Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.185147 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.188005 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.202052 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8wwq\" (UniqueName: \"kubernetes.io/projected/40ca9d77-9a9a-41dc-92eb-46706b9b3456-kube-api-access-v8wwq\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.202238 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ca9d77-9a9a-41dc-92eb-46706b9b3456-utilities\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.202323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5519f09-e6ba-4df8-8d28-d014f65fff6f-catalog-content\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.202540 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8744\" (UniqueName: \"kubernetes.io/projected/a5519f09-e6ba-4df8-8d28-d014f65fff6f-kube-api-access-b8744\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.202614 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ca9d77-9a9a-41dc-92eb-46706b9b3456-catalog-content\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.202697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5519f09-e6ba-4df8-8d28-d014f65fff6f-utilities\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.203216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5519f09-e6ba-4df8-8d28-d014f65fff6f-catalog-content\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.203441 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5519f09-e6ba-4df8-8d28-d014f65fff6f-utilities\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.205987 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5lhd"] Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.237964 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8744\" (UniqueName: \"kubernetes.io/projected/a5519f09-e6ba-4df8-8d28-d014f65fff6f-kube-api-access-b8744\") pod \"redhat-marketplace-wd2vs\" (UID: \"a5519f09-e6ba-4df8-8d28-d014f65fff6f\") " pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.303289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ca9d77-9a9a-41dc-92eb-46706b9b3456-catalog-content\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.303504 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8wwq\" (UniqueName: \"kubernetes.io/projected/40ca9d77-9a9a-41dc-92eb-46706b9b3456-kube-api-access-v8wwq\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.303552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ca9d77-9a9a-41dc-92eb-46706b9b3456-utilities\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.303941 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ca9d77-9a9a-41dc-92eb-46706b9b3456-catalog-content\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.304023 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ca9d77-9a9a-41dc-92eb-46706b9b3456-utilities\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.317756 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.326287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8wwq\" (UniqueName: \"kubernetes.io/projected/40ca9d77-9a9a-41dc-92eb-46706b9b3456-kube-api-access-v8wwq\") pod \"redhat-operators-g5lhd\" (UID: \"40ca9d77-9a9a-41dc-92eb-46706b9b3456\") " pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.499070 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:13 crc kubenswrapper[4744]: W1201 08:23:13.611169 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5519f09_e6ba_4df8_8d28_d014f65fff6f.slice/crio-7f7f0e9bb8fd5c2d37a9fcafbbf6f1ba9ff176ec5b3b6fff289b9d977fac2843 WatchSource:0}: Error finding container 7f7f0e9bb8fd5c2d37a9fcafbbf6f1ba9ff176ec5b3b6fff289b9d977fac2843: Status 404 returned error can't find the container with id 7f7f0e9bb8fd5c2d37a9fcafbbf6f1ba9ff176ec5b3b6fff289b9d977fac2843 Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.611532 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wd2vs"] Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.680497 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd2vs" event={"ID":"a5519f09-e6ba-4df8-8d28-d014f65fff6f","Type":"ContainerStarted","Data":"7f7f0e9bb8fd5c2d37a9fcafbbf6f1ba9ff176ec5b3b6fff289b9d977fac2843"} Dec 01 08:23:13 crc kubenswrapper[4744]: I1201 08:23:13.976310 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5lhd"] Dec 01 08:23:13 crc kubenswrapper[4744]: W1201 08:23:13.981225 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40ca9d77_9a9a_41dc_92eb_46706b9b3456.slice/crio-09129ea812ed06ce8ebbb9c43f3ab9aa5aaff6e150976692974fb28888e114cf WatchSource:0}: Error finding container 09129ea812ed06ce8ebbb9c43f3ab9aa5aaff6e150976692974fb28888e114cf: Status 404 returned error can't find the container with id 09129ea812ed06ce8ebbb9c43f3ab9aa5aaff6e150976692974fb28888e114cf Dec 01 08:23:14 crc kubenswrapper[4744]: I1201 08:23:14.688238 4744 generic.go:334] "Generic (PLEG): container finished" podID="a5519f09-e6ba-4df8-8d28-d014f65fff6f" containerID="c120b0640ea503fba106369a9a11e273bdc6360beef17c0216156c48802173d1" exitCode=0 Dec 01 08:23:14 crc kubenswrapper[4744]: I1201 08:23:14.688327 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd2vs" event={"ID":"a5519f09-e6ba-4df8-8d28-d014f65fff6f","Type":"ContainerDied","Data":"c120b0640ea503fba106369a9a11e273bdc6360beef17c0216156c48802173d1"} Dec 01 08:23:14 crc kubenswrapper[4744]: I1201 08:23:14.690567 4744 generic.go:334] "Generic (PLEG): container finished" podID="40ca9d77-9a9a-41dc-92eb-46706b9b3456" containerID="cfcc7557f3143667fe39a45f7523223f5ff4c5446f278817a10243e55f1d9370" exitCode=0 Dec 01 08:23:14 crc kubenswrapper[4744]: I1201 08:23:14.690620 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5lhd" event={"ID":"40ca9d77-9a9a-41dc-92eb-46706b9b3456","Type":"ContainerDied","Data":"cfcc7557f3143667fe39a45f7523223f5ff4c5446f278817a10243e55f1d9370"} Dec 01 08:23:14 crc kubenswrapper[4744]: I1201 08:23:14.690689 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5lhd" event={"ID":"40ca9d77-9a9a-41dc-92eb-46706b9b3456","Type":"ContainerStarted","Data":"09129ea812ed06ce8ebbb9c43f3ab9aa5aaff6e150976692974fb28888e114cf"} Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.384402 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m5kzr"] Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.386609 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.389882 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.405463 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5kzr"] Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.539289 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-utilities\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.539473 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-catalog-content\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.539559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m995w\" (UniqueName: \"kubernetes.io/projected/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-kube-api-access-m995w\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.588492 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x49gs"] Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.590928 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.594051 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.597565 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x49gs"] Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.641268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f8c61a-d957-4ad9-bdf9-16670c20020e-utilities\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.641367 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-725w5\" (UniqueName: \"kubernetes.io/projected/31f8c61a-d957-4ad9-bdf9-16670c20020e-kube-api-access-725w5\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.641446 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-utilities\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.641495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-catalog-content\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.641536 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m995w\" (UniqueName: \"kubernetes.io/projected/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-kube-api-access-m995w\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.641608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f8c61a-d957-4ad9-bdf9-16670c20020e-catalog-content\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.642450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-utilities\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.642899 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-catalog-content\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.675652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m995w\" (UniqueName: \"kubernetes.io/projected/bbeb5e92-32fb-4db8-80bf-e7617fdd4e67-kube-api-access-m995w\") pod \"certified-operators-m5kzr\" (UID: \"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67\") " pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.696744 4744 generic.go:334] "Generic (PLEG): container finished" podID="a5519f09-e6ba-4df8-8d28-d014f65fff6f" containerID="f626c482de2de9873f7ef4314dc89bf4c3f026f19474a55bfa2b6961aaa9c292" exitCode=0 Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.697576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd2vs" event={"ID":"a5519f09-e6ba-4df8-8d28-d014f65fff6f","Type":"ContainerDied","Data":"f626c482de2de9873f7ef4314dc89bf4c3f026f19474a55bfa2b6961aaa9c292"} Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.702302 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5lhd" event={"ID":"40ca9d77-9a9a-41dc-92eb-46706b9b3456","Type":"ContainerStarted","Data":"b7d4c24467c2a6e07d77bacbb680e3716ff7c08fb334012e7d2ae207e202811b"} Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.743143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f8c61a-d957-4ad9-bdf9-16670c20020e-catalog-content\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.743268 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f8c61a-d957-4ad9-bdf9-16670c20020e-utilities\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.743347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-725w5\" (UniqueName: \"kubernetes.io/projected/31f8c61a-d957-4ad9-bdf9-16670c20020e-kube-api-access-725w5\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.743909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f8c61a-d957-4ad9-bdf9-16670c20020e-utilities\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.744497 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f8c61a-d957-4ad9-bdf9-16670c20020e-catalog-content\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.762871 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-725w5\" (UniqueName: \"kubernetes.io/projected/31f8c61a-d957-4ad9-bdf9-16670c20020e-kube-api-access-725w5\") pod \"community-operators-x49gs\" (UID: \"31f8c61a-d957-4ad9-bdf9-16670c20020e\") " pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.776587 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:15 crc kubenswrapper[4744]: I1201 08:23:15.943852 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.008126 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5kzr"] Dec 01 08:23:16 crc kubenswrapper[4744]: W1201 08:23:16.017593 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbeb5e92_32fb_4db8_80bf_e7617fdd4e67.slice/crio-b007626b5884a519052ff3151895e386754edbdbd69fd796a287b54f24de4823 WatchSource:0}: Error finding container b007626b5884a519052ff3151895e386754edbdbd69fd796a287b54f24de4823: Status 404 returned error can't find the container with id b007626b5884a519052ff3151895e386754edbdbd69fd796a287b54f24de4823 Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.172561 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x49gs"] Dec 01 08:23:16 crc kubenswrapper[4744]: W1201 08:23:16.256902 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31f8c61a_d957_4ad9_bdf9_16670c20020e.slice/crio-f0ebba8c4738c2d5a1d078300e226589bb7089b906e7e13db5e70399a5876c85 WatchSource:0}: Error finding container f0ebba8c4738c2d5a1d078300e226589bb7089b906e7e13db5e70399a5876c85: Status 404 returned error can't find the container with id f0ebba8c4738c2d5a1d078300e226589bb7089b906e7e13db5e70399a5876c85 Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.709858 4744 generic.go:334] "Generic (PLEG): container finished" podID="31f8c61a-d957-4ad9-bdf9-16670c20020e" containerID="228efcdbbbe92acacb5b776720a0cb9aecf72db569ff18e36ef06829051e304e" exitCode=0 Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.710006 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x49gs" event={"ID":"31f8c61a-d957-4ad9-bdf9-16670c20020e","Type":"ContainerDied","Data":"228efcdbbbe92acacb5b776720a0cb9aecf72db569ff18e36ef06829051e304e"} Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.710070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x49gs" event={"ID":"31f8c61a-d957-4ad9-bdf9-16670c20020e","Type":"ContainerStarted","Data":"f0ebba8c4738c2d5a1d078300e226589bb7089b906e7e13db5e70399a5876c85"} Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.715457 4744 generic.go:334] "Generic (PLEG): container finished" podID="40ca9d77-9a9a-41dc-92eb-46706b9b3456" containerID="b7d4c24467c2a6e07d77bacbb680e3716ff7c08fb334012e7d2ae207e202811b" exitCode=0 Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.715591 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5lhd" event={"ID":"40ca9d77-9a9a-41dc-92eb-46706b9b3456","Type":"ContainerDied","Data":"b7d4c24467c2a6e07d77bacbb680e3716ff7c08fb334012e7d2ae207e202811b"} Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.725643 4744 generic.go:334] "Generic (PLEG): container finished" podID="bbeb5e92-32fb-4db8-80bf-e7617fdd4e67" containerID="4c7f51ce83977bd53ca0478a50a5c0eddcb6c8c23df47909ff8633de175d78fa" exitCode=0 Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.726157 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5kzr" event={"ID":"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67","Type":"ContainerDied","Data":"4c7f51ce83977bd53ca0478a50a5c0eddcb6c8c23df47909ff8633de175d78fa"} Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.726306 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5kzr" event={"ID":"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67","Type":"ContainerStarted","Data":"b007626b5884a519052ff3151895e386754edbdbd69fd796a287b54f24de4823"} Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.734482 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd2vs" event={"ID":"a5519f09-e6ba-4df8-8d28-d014f65fff6f","Type":"ContainerStarted","Data":"3717f23585661b7f1e4e2a49b5ead9e479ed3727ccc4cc87c79ef357a0c2e8d2"} Dec 01 08:23:16 crc kubenswrapper[4744]: I1201 08:23:16.759044 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wd2vs" podStartSLOduration=3.319088485 podStartE2EDuration="4.759024667s" podCreationTimestamp="2025-12-01 08:23:12 +0000 UTC" firstStartedPulling="2025-12-01 08:23:14.690097744 +0000 UTC m=+326.679155655" lastFinishedPulling="2025-12-01 08:23:16.130033916 +0000 UTC m=+328.119091837" observedRunningTime="2025-12-01 08:23:16.754614073 +0000 UTC m=+328.743671994" watchObservedRunningTime="2025-12-01 08:23:16.759024667 +0000 UTC m=+328.748082588" Dec 01 08:23:17 crc kubenswrapper[4744]: I1201 08:23:17.740301 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5kzr" event={"ID":"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67","Type":"ContainerStarted","Data":"5840f2e811a4a4ae58306eb161383ee936423af7dbf58f06980b5383029f7d02"} Dec 01 08:23:17 crc kubenswrapper[4744]: I1201 08:23:17.742738 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5lhd" event={"ID":"40ca9d77-9a9a-41dc-92eb-46706b9b3456","Type":"ContainerStarted","Data":"cba3cf66bd8bb6e13b38a7cf8c6462a9aa53898330c090ac31618595d5238c01"} Dec 01 08:23:17 crc kubenswrapper[4744]: I1201 08:23:17.782441 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5lhd" podStartSLOduration=2.159736126 podStartE2EDuration="4.782390415s" podCreationTimestamp="2025-12-01 08:23:13 +0000 UTC" firstStartedPulling="2025-12-01 08:23:14.692660117 +0000 UTC m=+326.681718048" lastFinishedPulling="2025-12-01 08:23:17.315314396 +0000 UTC m=+329.304372337" observedRunningTime="2025-12-01 08:23:17.782323943 +0000 UTC m=+329.771381864" watchObservedRunningTime="2025-12-01 08:23:17.782390415 +0000 UTC m=+329.771448336" Dec 01 08:23:18 crc kubenswrapper[4744]: I1201 08:23:18.755079 4744 generic.go:334] "Generic (PLEG): container finished" podID="bbeb5e92-32fb-4db8-80bf-e7617fdd4e67" containerID="5840f2e811a4a4ae58306eb161383ee936423af7dbf58f06980b5383029f7d02" exitCode=0 Dec 01 08:23:18 crc kubenswrapper[4744]: I1201 08:23:18.755164 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5kzr" event={"ID":"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67","Type":"ContainerDied","Data":"5840f2e811a4a4ae58306eb161383ee936423af7dbf58f06980b5383029f7d02"} Dec 01 08:23:18 crc kubenswrapper[4744]: I1201 08:23:18.760237 4744 generic.go:334] "Generic (PLEG): container finished" podID="31f8c61a-d957-4ad9-bdf9-16670c20020e" containerID="0d90e401cbdcce276d888f20487b90a67dbe7a9237dfd54764e8f7d7e0ff1f54" exitCode=0 Dec 01 08:23:18 crc kubenswrapper[4744]: I1201 08:23:18.761176 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x49gs" event={"ID":"31f8c61a-d957-4ad9-bdf9-16670c20020e","Type":"ContainerDied","Data":"0d90e401cbdcce276d888f20487b90a67dbe7a9237dfd54764e8f7d7e0ff1f54"} Dec 01 08:23:19 crc kubenswrapper[4744]: I1201 08:23:19.767038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x49gs" event={"ID":"31f8c61a-d957-4ad9-bdf9-16670c20020e","Type":"ContainerStarted","Data":"13a9eed7b44ff4801d6c791d95b914c6dbf73afe7b2e96068c71841f14c13066"} Dec 01 08:23:19 crc kubenswrapper[4744]: I1201 08:23:19.770959 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5kzr" event={"ID":"bbeb5e92-32fb-4db8-80bf-e7617fdd4e67","Type":"ContainerStarted","Data":"0da3ef3f5c4e7b8c91da7027545f6a234827ab6f3bef5f9602fabdf023264b28"} Dec 01 08:23:19 crc kubenswrapper[4744]: I1201 08:23:19.786076 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x49gs" podStartSLOduration=2.176768922 podStartE2EDuration="4.786058669s" podCreationTimestamp="2025-12-01 08:23:15 +0000 UTC" firstStartedPulling="2025-12-01 08:23:16.712633182 +0000 UTC m=+328.701691103" lastFinishedPulling="2025-12-01 08:23:19.321922909 +0000 UTC m=+331.310980850" observedRunningTime="2025-12-01 08:23:19.784634026 +0000 UTC m=+331.773691957" watchObservedRunningTime="2025-12-01 08:23:19.786058669 +0000 UTC m=+331.775116610" Dec 01 08:23:19 crc kubenswrapper[4744]: I1201 08:23:19.809764 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m5kzr" podStartSLOduration=2.260030723 podStartE2EDuration="4.809747636s" podCreationTimestamp="2025-12-01 08:23:15 +0000 UTC" firstStartedPulling="2025-12-01 08:23:16.728160012 +0000 UTC m=+328.717217923" lastFinishedPulling="2025-12-01 08:23:19.277876865 +0000 UTC m=+331.266934836" observedRunningTime="2025-12-01 08:23:19.807799417 +0000 UTC m=+331.796857348" watchObservedRunningTime="2025-12-01 08:23:19.809747636 +0000 UTC m=+331.798805567" Dec 01 08:23:23 crc kubenswrapper[4744]: I1201 08:23:23.318875 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:23 crc kubenswrapper[4744]: I1201 08:23:23.319378 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:23 crc kubenswrapper[4744]: I1201 08:23:23.373830 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:23 crc kubenswrapper[4744]: I1201 08:23:23.499460 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:23 crc kubenswrapper[4744]: I1201 08:23:23.499869 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:23 crc kubenswrapper[4744]: I1201 08:23:23.864570 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wd2vs" Dec 01 08:23:24 crc kubenswrapper[4744]: I1201 08:23:24.566776 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g5lhd" podUID="40ca9d77-9a9a-41dc-92eb-46706b9b3456" containerName="registry-server" probeResult="failure" output=< Dec 01 08:23:24 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 08:23:24 crc kubenswrapper[4744]: > Dec 01 08:23:25 crc kubenswrapper[4744]: I1201 08:23:25.777077 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:25 crc kubenswrapper[4744]: I1201 08:23:25.777163 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:25 crc kubenswrapper[4744]: I1201 08:23:25.846432 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:25 crc kubenswrapper[4744]: I1201 08:23:25.900241 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m5kzr" Dec 01 08:23:25 crc kubenswrapper[4744]: I1201 08:23:25.944536 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:25 crc kubenswrapper[4744]: I1201 08:23:25.944662 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:25 crc kubenswrapper[4744]: I1201 08:23:25.989346 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:26 crc kubenswrapper[4744]: I1201 08:23:26.870977 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x49gs" Dec 01 08:23:33 crc kubenswrapper[4744]: I1201 08:23:33.566284 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:33 crc kubenswrapper[4744]: I1201 08:23:33.636222 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5lhd" Dec 01 08:23:48 crc kubenswrapper[4744]: I1201 08:23:48.634539 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:23:48 crc kubenswrapper[4744]: I1201 08:23:48.635202 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.337756 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gb756"] Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.339827 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.349344 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gb756"] Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395065 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-registry-tls\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395165 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-trusted-ca\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395217 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-bound-sa-token\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395247 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395289 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rbmn\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-kube-api-access-5rbmn\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395379 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.395417 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-registry-certificates\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.421979 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.496545 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-bound-sa-token\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.496604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.496640 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rbmn\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-kube-api-access-5rbmn\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.496682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.496711 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-registry-certificates\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.496763 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-registry-tls\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.496801 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-trusted-ca\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.497211 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.498219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-trusted-ca\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.498225 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-registry-certificates\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.503862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.507970 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-registry-tls\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.511197 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-bound-sa-token\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.517063 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rbmn\" (UniqueName: \"kubernetes.io/projected/5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09-kube-api-access-5rbmn\") pod \"image-registry-66df7c8f76-gb756\" (UID: \"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09\") " pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:03 crc kubenswrapper[4744]: I1201 08:24:03.659703 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:04 crc kubenswrapper[4744]: I1201 08:24:04.140717 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gb756"] Dec 01 08:24:05 crc kubenswrapper[4744]: I1201 08:24:05.059158 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gb756" event={"ID":"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09","Type":"ContainerStarted","Data":"eb9452362bbd2baa934ae7403e33b87dfa629e547819312a6963d1e40a469130"} Dec 01 08:24:05 crc kubenswrapper[4744]: I1201 08:24:05.060192 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:05 crc kubenswrapper[4744]: I1201 08:24:05.060263 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gb756" event={"ID":"5ecaaf0b-1bbc-4fb4-944a-e1c426f50a09","Type":"ContainerStarted","Data":"c357c8fc47c3900f348e85b6e4b6be6821f93fe4f3612a2ef0a50fd2458885eb"} Dec 01 08:24:05 crc kubenswrapper[4744]: I1201 08:24:05.090771 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-gb756" podStartSLOduration=2.0907342 podStartE2EDuration="2.0907342s" podCreationTimestamp="2025-12-01 08:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:24:05.086241464 +0000 UTC m=+377.075299415" watchObservedRunningTime="2025-12-01 08:24:05.0907342 +0000 UTC m=+377.079792151" Dec 01 08:24:18 crc kubenswrapper[4744]: I1201 08:24:18.634077 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:24:18 crc kubenswrapper[4744]: I1201 08:24:18.635641 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:24:23 crc kubenswrapper[4744]: I1201 08:24:23.670434 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-gb756" Dec 01 08:24:23 crc kubenswrapper[4744]: I1201 08:24:23.741090 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l6djk"] Dec 01 08:24:27 crc kubenswrapper[4744]: I1201 08:24:27.665157 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9"] Dec 01 08:24:27 crc kubenswrapper[4744]: I1201 08:24:27.665742 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" podUID="2fc08073-673a-4dc6-a70e-20f850dbb988" containerName="route-controller-manager" containerID="cri-o://882618f84cf5a287981fbf6984d8baed0720154c2aef509511915689166d78f6" gracePeriod=30 Dec 01 08:24:28 crc kubenswrapper[4744]: I1201 08:24:28.194581 4744 generic.go:334] "Generic (PLEG): container finished" podID="2fc08073-673a-4dc6-a70e-20f850dbb988" containerID="882618f84cf5a287981fbf6984d8baed0720154c2aef509511915689166d78f6" exitCode=0 Dec 01 08:24:28 crc kubenswrapper[4744]: I1201 08:24:28.194687 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" event={"ID":"2fc08073-673a-4dc6-a70e-20f850dbb988","Type":"ContainerDied","Data":"882618f84cf5a287981fbf6984d8baed0720154c2aef509511915689166d78f6"} Dec 01 08:24:28 crc kubenswrapper[4744]: I1201 08:24:28.194801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" event={"ID":"2fc08073-673a-4dc6-a70e-20f850dbb988","Type":"ContainerDied","Data":"de9629e5f98640469a00587e513e056ef2e180ca7e0c96e1a4d9f28df6ed586d"} Dec 01 08:24:28 crc kubenswrapper[4744]: I1201 08:24:28.194820 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de9629e5f98640469a00587e513e056ef2e180ca7e0c96e1a4d9f28df6ed586d" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.078589 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.106052 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fc08073-673a-4dc6-a70e-20f850dbb988-serving-cert\") pod \"2fc08073-673a-4dc6-a70e-20f850dbb988\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.106119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khm88\" (UniqueName: \"kubernetes.io/projected/2fc08073-673a-4dc6-a70e-20f850dbb988-kube-api-access-khm88\") pod \"2fc08073-673a-4dc6-a70e-20f850dbb988\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.106151 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-config\") pod \"2fc08073-673a-4dc6-a70e-20f850dbb988\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.106215 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-client-ca\") pod \"2fc08073-673a-4dc6-a70e-20f850dbb988\" (UID: \"2fc08073-673a-4dc6-a70e-20f850dbb988\") " Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.107266 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-config" (OuterVolumeSpecName: "config") pod "2fc08073-673a-4dc6-a70e-20f850dbb988" (UID: "2fc08073-673a-4dc6-a70e-20f850dbb988"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.107743 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-client-ca" (OuterVolumeSpecName: "client-ca") pod "2fc08073-673a-4dc6-a70e-20f850dbb988" (UID: "2fc08073-673a-4dc6-a70e-20f850dbb988"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.113683 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc08073-673a-4dc6-a70e-20f850dbb988-kube-api-access-khm88" (OuterVolumeSpecName: "kube-api-access-khm88") pod "2fc08073-673a-4dc6-a70e-20f850dbb988" (UID: "2fc08073-673a-4dc6-a70e-20f850dbb988"). InnerVolumeSpecName "kube-api-access-khm88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.114239 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc08073-673a-4dc6-a70e-20f850dbb988-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2fc08073-673a-4dc6-a70e-20f850dbb988" (UID: "2fc08073-673a-4dc6-a70e-20f850dbb988"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.138505 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2"] Dec 01 08:24:29 crc kubenswrapper[4744]: E1201 08:24:29.138804 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc08073-673a-4dc6-a70e-20f850dbb988" containerName="route-controller-manager" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.138825 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc08073-673a-4dc6-a70e-20f850dbb988" containerName="route-controller-manager" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.138970 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc08073-673a-4dc6-a70e-20f850dbb988" containerName="route-controller-manager" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.139442 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.141581 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2"] Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.198668 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.207031 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.207060 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fc08073-673a-4dc6-a70e-20f850dbb988-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.207074 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khm88\" (UniqueName: \"kubernetes.io/projected/2fc08073-673a-4dc6-a70e-20f850dbb988-kube-api-access-khm88\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.207087 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc08073-673a-4dc6-a70e-20f850dbb988-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.222798 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9"] Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.231604 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78d46db78f-zt9b9"] Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.308067 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-serving-cert\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.308130 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-client-ca\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.308152 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2zwq\" (UniqueName: \"kubernetes.io/projected/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-kube-api-access-c2zwq\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.308364 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-config\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.410122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-config\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.410175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-serving-cert\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.410207 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-client-ca\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.410227 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2zwq\" (UniqueName: \"kubernetes.io/projected/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-kube-api-access-c2zwq\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.411347 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-client-ca\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.411969 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-config\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.416111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-serving-cert\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.429965 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2zwq\" (UniqueName: \"kubernetes.io/projected/b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9-kube-api-access-c2zwq\") pod \"route-controller-manager-f9988df89-jxbx2\" (UID: \"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9\") " pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.456075 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:29 crc kubenswrapper[4744]: I1201 08:24:29.883232 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2"] Dec 01 08:24:30 crc kubenswrapper[4744]: I1201 08:24:30.208640 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" event={"ID":"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9","Type":"ContainerStarted","Data":"12e1d309de5b9fbe7becc675ec82bb1af166833231fbfaf3c51dbe88c4dccdb9"} Dec 01 08:24:30 crc kubenswrapper[4744]: I1201 08:24:30.209081 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:30 crc kubenswrapper[4744]: I1201 08:24:30.209104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" event={"ID":"b2a60bdc-4dfc-4ca5-a59e-af77ccc6fae9","Type":"ContainerStarted","Data":"66c7630706e4d148c4552a5e23f8fe75a9b92c5c0109897d66092889def8802d"} Dec 01 08:24:30 crc kubenswrapper[4744]: I1201 08:24:30.232914 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" podStartSLOduration=3.2328888669999998 podStartE2EDuration="3.232888867s" podCreationTimestamp="2025-12-01 08:24:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:24:30.226801543 +0000 UTC m=+402.215859474" watchObservedRunningTime="2025-12-01 08:24:30.232888867 +0000 UTC m=+402.221946838" Dec 01 08:24:30 crc kubenswrapper[4744]: I1201 08:24:30.294847 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fc08073-673a-4dc6-a70e-20f850dbb988" path="/var/lib/kubelet/pods/2fc08073-673a-4dc6-a70e-20f850dbb988/volumes" Dec 01 08:24:30 crc kubenswrapper[4744]: I1201 08:24:30.644257 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-f9988df89-jxbx2" Dec 01 08:24:48 crc kubenswrapper[4744]: I1201 08:24:48.645720 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:24:48 crc kubenswrapper[4744]: I1201 08:24:48.646446 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:24:48 crc kubenswrapper[4744]: I1201 08:24:48.646537 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:24:48 crc kubenswrapper[4744]: I1201 08:24:48.647363 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b444480d1eb4ea273efbce3cf9243860be3f4dff96451c41a797893d150f669"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:24:48 crc kubenswrapper[4744]: I1201 08:24:48.647490 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://4b444480d1eb4ea273efbce3cf9243860be3f4dff96451c41a797893d150f669" gracePeriod=600 Dec 01 08:24:48 crc kubenswrapper[4744]: I1201 08:24:48.792455 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" podUID="5516996f-81ea-44d0-b5a1-5af61f00af8c" containerName="registry" containerID="cri-o://3ba489857f29415fc09c5af9fb635237717ee48896e20f7fca1b3e5e69eca709" gracePeriod=30 Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.360554 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="4b444480d1eb4ea273efbce3cf9243860be3f4dff96451c41a797893d150f669" exitCode=0 Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.360678 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"4b444480d1eb4ea273efbce3cf9243860be3f4dff96451c41a797893d150f669"} Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.360736 4744 scope.go:117] "RemoveContainer" containerID="8e631745a775bba6e3becbeba8765f0b08dc8458c8eb0b86feb7fb1abbfbbb19" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.371715 4744 generic.go:334] "Generic (PLEG): container finished" podID="5516996f-81ea-44d0-b5a1-5af61f00af8c" containerID="3ba489857f29415fc09c5af9fb635237717ee48896e20f7fca1b3e5e69eca709" exitCode=0 Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.371837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" event={"ID":"5516996f-81ea-44d0-b5a1-5af61f00af8c","Type":"ContainerDied","Data":"3ba489857f29415fc09c5af9fb635237717ee48896e20f7fca1b3e5e69eca709"} Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.648647 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.733825 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-certificates\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.733876 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-tls\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.733906 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5516996f-81ea-44d0-b5a1-5af61f00af8c-installation-pull-secrets\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.733969 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5516996f-81ea-44d0-b5a1-5af61f00af8c-ca-trust-extracted\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.734017 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-trusted-ca\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.734037 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-bound-sa-token\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.734054 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t85fp\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-kube-api-access-t85fp\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.734177 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"5516996f-81ea-44d0-b5a1-5af61f00af8c\" (UID: \"5516996f-81ea-44d0-b5a1-5af61f00af8c\") " Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.734715 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.734771 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.739499 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5516996f-81ea-44d0-b5a1-5af61f00af8c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.740054 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-kube-api-access-t85fp" (OuterVolumeSpecName: "kube-api-access-t85fp") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "kube-api-access-t85fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.740242 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.742151 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.744708 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.753584 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5516996f-81ea-44d0-b5a1-5af61f00af8c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "5516996f-81ea-44d0-b5a1-5af61f00af8c" (UID: "5516996f-81ea-44d0-b5a1-5af61f00af8c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.835007 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.835044 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.835055 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t85fp\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-kube-api-access-t85fp\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.835065 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.835073 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5516996f-81ea-44d0-b5a1-5af61f00af8c-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.835081 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5516996f-81ea-44d0-b5a1-5af61f00af8c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:49 crc kubenswrapper[4744]: I1201 08:24:49.835089 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5516996f-81ea-44d0-b5a1-5af61f00af8c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 08:24:50 crc kubenswrapper[4744]: I1201 08:24:50.381129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"08978ca5ab125cd2f75e7461dc8ec1a4a02dd4cdb1c39a55fa75057d29c93966"} Dec 01 08:24:50 crc kubenswrapper[4744]: I1201 08:24:50.384616 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" event={"ID":"5516996f-81ea-44d0-b5a1-5af61f00af8c","Type":"ContainerDied","Data":"d90e8b978ecf142bc1d3be88896d315f65bc8d44faf70adb7f33710cceeab76c"} Dec 01 08:24:50 crc kubenswrapper[4744]: I1201 08:24:50.384649 4744 scope.go:117] "RemoveContainer" containerID="3ba489857f29415fc09c5af9fb635237717ee48896e20f7fca1b3e5e69eca709" Dec 01 08:24:50 crc kubenswrapper[4744]: I1201 08:24:50.384792 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l6djk" Dec 01 08:24:50 crc kubenswrapper[4744]: I1201 08:24:50.428488 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l6djk"] Dec 01 08:24:50 crc kubenswrapper[4744]: I1201 08:24:50.433550 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l6djk"] Dec 01 08:24:52 crc kubenswrapper[4744]: I1201 08:24:52.297004 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5516996f-81ea-44d0-b5a1-5af61f00af8c" path="/var/lib/kubelet/pods/5516996f-81ea-44d0-b5a1-5af61f00af8c/volumes" Dec 01 08:27:18 crc kubenswrapper[4744]: I1201 08:27:18.634915 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:27:18 crc kubenswrapper[4744]: I1201 08:27:18.635565 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:27:48 crc kubenswrapper[4744]: I1201 08:27:48.634550 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:27:48 crc kubenswrapper[4744]: I1201 08:27:48.635121 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.351599 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bp2pq"] Dec 01 08:28:13 crc kubenswrapper[4744]: E1201 08:28:13.352596 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5516996f-81ea-44d0-b5a1-5af61f00af8c" containerName="registry" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.352622 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5516996f-81ea-44d0-b5a1-5af61f00af8c" containerName="registry" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.353123 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5516996f-81ea-44d0-b5a1-5af61f00af8c" containerName="registry" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.353743 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.357685 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.357961 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.358293 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wrw5b" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.364306 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bjbkb"] Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.365032 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bjbkb" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.367568 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-9gdjh" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.368473 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-w864b"] Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.369044 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.371476 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jrzjp" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.380299 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bp2pq"] Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.387265 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bjbkb"] Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.391137 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-w864b"] Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.446279 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfbnc\" (UniqueName: \"kubernetes.io/projected/e05026bf-1db3-4a62-baff-a20dd3d8fd65-kube-api-access-nfbnc\") pod \"cert-manager-5b446d88c5-bjbkb\" (UID: \"e05026bf-1db3-4a62-baff-a20dd3d8fd65\") " pod="cert-manager/cert-manager-5b446d88c5-bjbkb" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.446349 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qqrq\" (UniqueName: \"kubernetes.io/projected/5bc4fc9c-b3ed-40f1-b951-8587803ba399-kube-api-access-5qqrq\") pod \"cert-manager-cainjector-7f985d654d-bp2pq\" (UID: \"5bc4fc9c-b3ed-40f1-b951-8587803ba399\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.547449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfbnc\" (UniqueName: \"kubernetes.io/projected/e05026bf-1db3-4a62-baff-a20dd3d8fd65-kube-api-access-nfbnc\") pod \"cert-manager-5b446d88c5-bjbkb\" (UID: \"e05026bf-1db3-4a62-baff-a20dd3d8fd65\") " pod="cert-manager/cert-manager-5b446d88c5-bjbkb" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.547516 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qqrq\" (UniqueName: \"kubernetes.io/projected/5bc4fc9c-b3ed-40f1-b951-8587803ba399-kube-api-access-5qqrq\") pod \"cert-manager-cainjector-7f985d654d-bp2pq\" (UID: \"5bc4fc9c-b3ed-40f1-b951-8587803ba399\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.547614 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frsrq\" (UniqueName: \"kubernetes.io/projected/db8c6300-bdd5-4fe7-af23-ed9feb32b844-kube-api-access-frsrq\") pod \"cert-manager-webhook-5655c58dd6-w864b\" (UID: \"db8c6300-bdd5-4fe7-af23-ed9feb32b844\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.565844 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfbnc\" (UniqueName: \"kubernetes.io/projected/e05026bf-1db3-4a62-baff-a20dd3d8fd65-kube-api-access-nfbnc\") pod \"cert-manager-5b446d88c5-bjbkb\" (UID: \"e05026bf-1db3-4a62-baff-a20dd3d8fd65\") " pod="cert-manager/cert-manager-5b446d88c5-bjbkb" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.567646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qqrq\" (UniqueName: \"kubernetes.io/projected/5bc4fc9c-b3ed-40f1-b951-8587803ba399-kube-api-access-5qqrq\") pod \"cert-manager-cainjector-7f985d654d-bp2pq\" (UID: \"5bc4fc9c-b3ed-40f1-b951-8587803ba399\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.649033 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frsrq\" (UniqueName: \"kubernetes.io/projected/db8c6300-bdd5-4fe7-af23-ed9feb32b844-kube-api-access-frsrq\") pod \"cert-manager-webhook-5655c58dd6-w864b\" (UID: \"db8c6300-bdd5-4fe7-af23-ed9feb32b844\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.667518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frsrq\" (UniqueName: \"kubernetes.io/projected/db8c6300-bdd5-4fe7-af23-ed9feb32b844-kube-api-access-frsrq\") pod \"cert-manager-webhook-5655c58dd6-w864b\" (UID: \"db8c6300-bdd5-4fe7-af23-ed9feb32b844\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.674699 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.682075 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bjbkb" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.688084 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.972258 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-w864b"] Dec 01 08:28:13 crc kubenswrapper[4744]: I1201 08:28:13.989067 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:28:14 crc kubenswrapper[4744]: I1201 08:28:14.136347 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bjbkb"] Dec 01 08:28:14 crc kubenswrapper[4744]: I1201 08:28:14.142548 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bp2pq"] Dec 01 08:28:14 crc kubenswrapper[4744]: W1201 08:28:14.143574 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode05026bf_1db3_4a62_baff_a20dd3d8fd65.slice/crio-13cf8978a626c4eac7fa248634f31eb09dfcd4824026ab2404e8be48862abfcf WatchSource:0}: Error finding container 13cf8978a626c4eac7fa248634f31eb09dfcd4824026ab2404e8be48862abfcf: Status 404 returned error can't find the container with id 13cf8978a626c4eac7fa248634f31eb09dfcd4824026ab2404e8be48862abfcf Dec 01 08:28:14 crc kubenswrapper[4744]: W1201 08:28:14.146691 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bc4fc9c_b3ed_40f1_b951_8587803ba399.slice/crio-e2f6fc35d5c035cfffeb8e968136874a01dd8aab9d88bf51c371aa0bfafefef7 WatchSource:0}: Error finding container e2f6fc35d5c035cfffeb8e968136874a01dd8aab9d88bf51c371aa0bfafefef7: Status 404 returned error can't find the container with id e2f6fc35d5c035cfffeb8e968136874a01dd8aab9d88bf51c371aa0bfafefef7 Dec 01 08:28:14 crc kubenswrapper[4744]: I1201 08:28:14.199774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bjbkb" event={"ID":"e05026bf-1db3-4a62-baff-a20dd3d8fd65","Type":"ContainerStarted","Data":"13cf8978a626c4eac7fa248634f31eb09dfcd4824026ab2404e8be48862abfcf"} Dec 01 08:28:14 crc kubenswrapper[4744]: I1201 08:28:14.200822 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" event={"ID":"db8c6300-bdd5-4fe7-af23-ed9feb32b844","Type":"ContainerStarted","Data":"6e18f4998e1be9bcf3b078aa65709f1e8fc1f451e90e68115f6004d5f39e7937"} Dec 01 08:28:14 crc kubenswrapper[4744]: I1201 08:28:14.201848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" event={"ID":"5bc4fc9c-b3ed-40f1-b951-8587803ba399","Type":"ContainerStarted","Data":"e2f6fc35d5c035cfffeb8e968136874a01dd8aab9d88bf51c371aa0bfafefef7"} Dec 01 08:28:16 crc kubenswrapper[4744]: I1201 08:28:16.217100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" event={"ID":"db8c6300-bdd5-4fe7-af23-ed9feb32b844","Type":"ContainerStarted","Data":"3fe7150ca02fa88edbcec0e4dd92cba1685ad1a8fd691e292312ddbed0a0be77"} Dec 01 08:28:16 crc kubenswrapper[4744]: I1201 08:28:16.218115 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" Dec 01 08:28:16 crc kubenswrapper[4744]: I1201 08:28:16.234489 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" podStartSLOduration=1.334462093 podStartE2EDuration="3.234468109s" podCreationTimestamp="2025-12-01 08:28:13 +0000 UTC" firstStartedPulling="2025-12-01 08:28:13.988777047 +0000 UTC m=+625.977834978" lastFinishedPulling="2025-12-01 08:28:15.888783033 +0000 UTC m=+627.877840994" observedRunningTime="2025-12-01 08:28:16.232755633 +0000 UTC m=+628.221813554" watchObservedRunningTime="2025-12-01 08:28:16.234468109 +0000 UTC m=+628.223526030" Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.230593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" event={"ID":"5bc4fc9c-b3ed-40f1-b951-8587803ba399","Type":"ContainerStarted","Data":"d6df5b9f7ac491e646caf120a9bbb1d75f00553e9363f1255e1447cf8b90b403"} Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.233067 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bjbkb" event={"ID":"e05026bf-1db3-4a62-baff-a20dd3d8fd65","Type":"ContainerStarted","Data":"0dc87fb5b68141e75adc76055e2515e1e6654441ef6fa1170db6d131c1d4776b"} Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.251615 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-bp2pq" podStartSLOduration=1.981028158 podStartE2EDuration="5.25158804s" podCreationTimestamp="2025-12-01 08:28:13 +0000 UTC" firstStartedPulling="2025-12-01 08:28:14.148490547 +0000 UTC m=+626.137548478" lastFinishedPulling="2025-12-01 08:28:17.419050429 +0000 UTC m=+629.408108360" observedRunningTime="2025-12-01 08:28:18.247820669 +0000 UTC m=+630.236878650" watchObservedRunningTime="2025-12-01 08:28:18.25158804 +0000 UTC m=+630.240645991" Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.271870 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-bjbkb" podStartSLOduration=1.9925214869999999 podStartE2EDuration="5.271841884s" podCreationTimestamp="2025-12-01 08:28:13 +0000 UTC" firstStartedPulling="2025-12-01 08:28:14.145820915 +0000 UTC m=+626.134878846" lastFinishedPulling="2025-12-01 08:28:17.425141322 +0000 UTC m=+629.414199243" observedRunningTime="2025-12-01 08:28:18.26756394 +0000 UTC m=+630.256621871" watchObservedRunningTime="2025-12-01 08:28:18.271841884 +0000 UTC m=+630.260899845" Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.634372 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.634512 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.634576 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.635402 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"08978ca5ab125cd2f75e7461dc8ec1a4a02dd4cdb1c39a55fa75057d29c93966"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:28:18 crc kubenswrapper[4744]: I1201 08:28:18.635547 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://08978ca5ab125cd2f75e7461dc8ec1a4a02dd4cdb1c39a55fa75057d29c93966" gracePeriod=600 Dec 01 08:28:19 crc kubenswrapper[4744]: I1201 08:28:19.242295 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="08978ca5ab125cd2f75e7461dc8ec1a4a02dd4cdb1c39a55fa75057d29c93966" exitCode=0 Dec 01 08:28:19 crc kubenswrapper[4744]: I1201 08:28:19.242463 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"08978ca5ab125cd2f75e7461dc8ec1a4a02dd4cdb1c39a55fa75057d29c93966"} Dec 01 08:28:19 crc kubenswrapper[4744]: I1201 08:28:19.242615 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"0d6c923f3edaf0f41eb1f1e74b19db369be4219783160ca17cb5ceacdf850175"} Dec 01 08:28:19 crc kubenswrapper[4744]: I1201 08:28:19.242648 4744 scope.go:117] "RemoveContainer" containerID="4b444480d1eb4ea273efbce3cf9243860be3f4dff96451c41a797893d150f669" Dec 01 08:28:23 crc kubenswrapper[4744]: I1201 08:28:23.693038 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-w864b" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.107538 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2sflk"] Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.108324 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-controller" containerID="cri-o://afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.108429 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="northd" containerID="cri-o://d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.108480 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="sbdb" containerID="cri-o://336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.108472 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-acl-logging" containerID="cri-o://bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.108449 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-node" containerID="cri-o://d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.108610 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="nbdb" containerID="cri-o://e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.108748 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.141200 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" containerID="cri-o://634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b" gracePeriod=30 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.282115 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovnkube-controller/3.log" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.284238 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovn-acl-logging/0.log" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.284702 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovn-controller/0.log" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.285024 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b" exitCode=0 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.285053 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22" exitCode=0 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.285060 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8" exitCode=0 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.285068 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626" exitCode=0 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.285074 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff" exitCode=143 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.285081 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f" exitCode=143 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.286972 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/2.log" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.287396 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/1.log" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.287476 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a5e5532-8675-4f3f-aa30-7c22bc95b311" containerID="c5cd4b8744689d1838bd2d0e5d0166485469706e4e1b6d8f26489c876c06341e" exitCode=2 Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292736 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b"} Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292791 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22"} Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292806 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8"} Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626"} Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292831 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff"} Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f"} Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292854 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerDied","Data":"c5cd4b8744689d1838bd2d0e5d0166485469706e4e1b6d8f26489c876c06341e"} Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.292960 4744 scope.go:117] "RemoveContainer" containerID="7c46d8b2debca88d1c4a3adc1004776b52c6b3e7af5f326d5696c01031a052d1" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.293330 4744 scope.go:117] "RemoveContainer" containerID="c5cd4b8744689d1838bd2d0e5d0166485469706e4e1b6d8f26489c876c06341e" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.294590 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vsh4r_openshift-multus(6a5e5532-8675-4f3f-aa30-7c22bc95b311)\"" pod="openshift-multus/multus-vsh4r" podUID="6a5e5532-8675-4f3f-aa30-7c22bc95b311" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.401086 4744 scope.go:117] "RemoveContainer" containerID="e747354416553f0f3e4c681570dc7236c6ac67b2ef7f4661c7054b74011f8639" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.410040 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovn-acl-logging/0.log" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.410994 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovn-controller/0.log" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.411846 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474064 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hh8pr"] Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474307 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474324 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474338 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-node" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474346 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-node" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474358 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474366 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474379 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="sbdb" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474386 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="sbdb" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474402 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="northd" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474432 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="northd" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474442 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-acl-logging" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474450 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-acl-logging" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="nbdb" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474467 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="nbdb" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474475 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474482 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474494 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474502 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474511 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kubecfg-setup" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474519 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kubecfg-setup" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474531 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474538 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474550 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474558 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474664 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="nbdb" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474674 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474683 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474697 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474754 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474771 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="kube-rbac-proxy-node" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474782 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474794 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="northd" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474804 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="sbdb" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474817 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovn-acl-logging" Dec 01 08:28:24 crc kubenswrapper[4744]: E1201 08:28:24.474964 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.474977 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.475117 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.475455 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerName="ovnkube-controller" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.477356 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.494951 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-bin\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495043 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-systemd\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495098 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-ovn\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-env-overrides\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495145 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-log-socket\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495160 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-slash\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495177 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-systemd-units\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495192 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-netns\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-kubelet\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495233 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-config\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495253 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1df124eb-0511-4702-85d8-3e324c59bb0d-ovn-node-metrics-cert\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495274 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-etc-openvswitch\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495293 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-openvswitch\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495322 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxzmp\" (UniqueName: \"kubernetes.io/projected/1df124eb-0511-4702-85d8-3e324c59bb0d-kube-api-access-hxzmp\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495351 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-var-lib-openvswitch\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-slash" (OuterVolumeSpecName: "host-slash") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495369 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495387 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-node-log\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495392 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495424 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-script-lib\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495483 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-netd\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495513 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-ovn-kubernetes\") pod \"1df124eb-0511-4702-85d8-3e324c59bb0d\" (UID: \"1df124eb-0511-4702-85d8-3e324c59bb0d\") " Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495799 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495831 4744 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495658 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-log-socket" (OuterVolumeSpecName: "log-socket") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495651 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495693 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495712 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495729 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495736 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495770 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495787 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495866 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495866 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-node-log" (OuterVolumeSpecName: "node-log") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.495888 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.496171 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.496535 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.496644 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.502487 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1df124eb-0511-4702-85d8-3e324c59bb0d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.503499 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1df124eb-0511-4702-85d8-3e324c59bb0d-kube-api-access-hxzmp" (OuterVolumeSpecName: "kube-api-access-hxzmp") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "kube-api-access-hxzmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.520595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1df124eb-0511-4702-85d8-3e324c59bb0d" (UID: "1df124eb-0511-4702-85d8-3e324c59bb0d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597287 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-ovnkube-config\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597360 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-run-ovn-kubernetes\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597399 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-log-socket\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-ovnkube-script-lib\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597569 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-var-lib-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597656 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597692 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-cni-netd\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597729 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28e32de2-c0ce-426c-b30e-215073fba2bf-ovn-node-metrics-cert\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597770 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-slash\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597830 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-ovn\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597868 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-systemd-units\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597909 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plrfx\" (UniqueName: \"kubernetes.io/projected/28e32de2-c0ce-426c-b30e-215073fba2bf-kube-api-access-plrfx\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597961 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-env-overrides\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.597996 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-cni-bin\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598026 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-kubelet\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-run-netns\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-etc-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598234 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-systemd\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598270 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-node-log\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598314 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598472 4744 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598495 4744 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598512 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598527 4744 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598543 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598559 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1df124eb-0511-4702-85d8-3e324c59bb0d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598574 4744 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598589 4744 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598606 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxzmp\" (UniqueName: \"kubernetes.io/projected/1df124eb-0511-4702-85d8-3e324c59bb0d-kube-api-access-hxzmp\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598622 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598639 4744 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598655 4744 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598674 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598713 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598731 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598746 4744 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598759 4744 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1df124eb-0511-4702-85d8-3e324c59bb0d-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.598773 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1df124eb-0511-4702-85d8-3e324c59bb0d-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.699923 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-cni-netd\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700044 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28e32de2-c0ce-426c-b30e-215073fba2bf-ovn-node-metrics-cert\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700074 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-slash\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-ovn\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-systemd-units\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700165 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plrfx\" (UniqueName: \"kubernetes.io/projected/28e32de2-c0ce-426c-b30e-215073fba2bf-kube-api-access-plrfx\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-env-overrides\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-cni-bin\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-kubelet\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700290 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-run-netns\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700324 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-etc-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-systemd\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700426 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-node-log\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700551 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-ovnkube-config\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700583 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-run-ovn-kubernetes\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700612 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-log-socket\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700644 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-ovnkube-script-lib\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700672 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-var-lib-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-var-lib-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700831 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.700873 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-cni-netd\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702570 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-run-netns\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702570 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702636 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-etc-openvswitch\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702718 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-node-log\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702769 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-kubelet\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702775 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-systemd\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702821 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-run-ovn\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-log-socket\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702816 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-run-ovn-kubernetes\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702861 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-slash\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-host-cni-bin\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.702945 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28e32de2-c0ce-426c-b30e-215073fba2bf-systemd-units\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.703693 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-env-overrides\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.707587 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-ovnkube-config\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.709533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28e32de2-c0ce-426c-b30e-215073fba2bf-ovnkube-script-lib\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.711204 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28e32de2-c0ce-426c-b30e-215073fba2bf-ovn-node-metrics-cert\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.730455 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plrfx\" (UniqueName: \"kubernetes.io/projected/28e32de2-c0ce-426c-b30e-215073fba2bf-kube-api-access-plrfx\") pod \"ovnkube-node-hh8pr\" (UID: \"28e32de2-c0ce-426c-b30e-215073fba2bf\") " pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:24 crc kubenswrapper[4744]: I1201 08:28:24.796335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.299789 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/2.log" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.302962 4744 generic.go:334] "Generic (PLEG): container finished" podID="28e32de2-c0ce-426c-b30e-215073fba2bf" containerID="24a8ae0d3b2bd0bc38995a1fcdf0668f9add661ce524f200276aa0c48b16f6a3" exitCode=0 Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.303082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerDied","Data":"24a8ae0d3b2bd0bc38995a1fcdf0668f9add661ce524f200276aa0c48b16f6a3"} Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.303137 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"5e24740d124a7ce5f82dc5e871c80210a84521b4da3bb4f33b1ff1bd8daa1d48"} Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.311778 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovn-acl-logging/0.log" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.312615 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2sflk_1df124eb-0511-4702-85d8-3e324c59bb0d/ovn-controller/0.log" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.313222 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1" exitCode=0 Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.313301 4744 generic.go:334] "Generic (PLEG): container finished" podID="1df124eb-0511-4702-85d8-3e324c59bb0d" containerID="d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed" exitCode=0 Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.313342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1"} Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.313386 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed"} Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.313432 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" event={"ID":"1df124eb-0511-4702-85d8-3e324c59bb0d","Type":"ContainerDied","Data":"e4f00f2099e951bf715b1b3196e6597595c7cefbabb3eb9f71ccbaa0ef227d6c"} Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.313446 4744 scope.go:117] "RemoveContainer" containerID="634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.313456 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2sflk" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.342576 4744 scope.go:117] "RemoveContainer" containerID="336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.398373 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2sflk"] Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.410431 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2sflk"] Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.415591 4744 scope.go:117] "RemoveContainer" containerID="e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.437467 4744 scope.go:117] "RemoveContainer" containerID="d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.465568 4744 scope.go:117] "RemoveContainer" containerID="ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.493940 4744 scope.go:117] "RemoveContainer" containerID="d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.515370 4744 scope.go:117] "RemoveContainer" containerID="bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.529729 4744 scope.go:117] "RemoveContainer" containerID="afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.557978 4744 scope.go:117] "RemoveContainer" containerID="5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.577582 4744 scope.go:117] "RemoveContainer" containerID="634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.577980 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b\": container with ID starting with 634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b not found: ID does not exist" containerID="634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.578022 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b"} err="failed to get container status \"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b\": rpc error: code = NotFound desc = could not find container \"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b\": container with ID starting with 634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.578050 4744 scope.go:117] "RemoveContainer" containerID="336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.578501 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\": container with ID starting with 336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1 not found: ID does not exist" containerID="336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.578518 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1"} err="failed to get container status \"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\": rpc error: code = NotFound desc = could not find container \"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\": container with ID starting with 336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.578531 4744 scope.go:117] "RemoveContainer" containerID="e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.578855 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\": container with ID starting with e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22 not found: ID does not exist" containerID="e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.578879 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22"} err="failed to get container status \"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\": rpc error: code = NotFound desc = could not find container \"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\": container with ID starting with e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.578900 4744 scope.go:117] "RemoveContainer" containerID="d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.579224 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\": container with ID starting with d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed not found: ID does not exist" containerID="d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.579288 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed"} err="failed to get container status \"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\": rpc error: code = NotFound desc = could not find container \"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\": container with ID starting with d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.579345 4744 scope.go:117] "RemoveContainer" containerID="ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.579828 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\": container with ID starting with ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8 not found: ID does not exist" containerID="ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.579847 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8"} err="failed to get container status \"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\": rpc error: code = NotFound desc = could not find container \"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\": container with ID starting with ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.579860 4744 scope.go:117] "RemoveContainer" containerID="d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.580070 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\": container with ID starting with d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626 not found: ID does not exist" containerID="d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.580095 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626"} err="failed to get container status \"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\": rpc error: code = NotFound desc = could not find container \"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\": container with ID starting with d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.580111 4744 scope.go:117] "RemoveContainer" containerID="bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.580384 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\": container with ID starting with bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff not found: ID does not exist" containerID="bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.580505 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff"} err="failed to get container status \"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\": rpc error: code = NotFound desc = could not find container \"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\": container with ID starting with bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.580544 4744 scope.go:117] "RemoveContainer" containerID="afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.580852 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\": container with ID starting with afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f not found: ID does not exist" containerID="afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.580882 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f"} err="failed to get container status \"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\": rpc error: code = NotFound desc = could not find container \"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\": container with ID starting with afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.580902 4744 scope.go:117] "RemoveContainer" containerID="5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6" Dec 01 08:28:25 crc kubenswrapper[4744]: E1201 08:28:25.581167 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\": container with ID starting with 5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6 not found: ID does not exist" containerID="5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.581204 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6"} err="failed to get container status \"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\": rpc error: code = NotFound desc = could not find container \"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\": container with ID starting with 5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.581240 4744 scope.go:117] "RemoveContainer" containerID="634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.581530 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b"} err="failed to get container status \"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b\": rpc error: code = NotFound desc = could not find container \"634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b\": container with ID starting with 634f60aedb43fcb8de0b2f33ca007c69d5ce32d2f3c29be025ab66655905e83b not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.581561 4744 scope.go:117] "RemoveContainer" containerID="336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.581817 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1"} err="failed to get container status \"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\": rpc error: code = NotFound desc = could not find container \"336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1\": container with ID starting with 336d788b6b8226e1c5cc46085c0e48ab65927e42d5c27f487aab3fd50743f7d1 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.581838 4744 scope.go:117] "RemoveContainer" containerID="e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582093 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22"} err="failed to get container status \"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\": rpc error: code = NotFound desc = could not find container \"e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22\": container with ID starting with e0f8a0a88889acaf3710af11b0c79857ff09964c1f6b90454b6f63604504db22 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582117 4744 scope.go:117] "RemoveContainer" containerID="d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582339 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed"} err="failed to get container status \"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\": rpc error: code = NotFound desc = could not find container \"d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed\": container with ID starting with d84ced5115c3e4a1fc56612add92456bb35acd9841e840dc8e64cb3281ebb4ed not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582354 4744 scope.go:117] "RemoveContainer" containerID="ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582596 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8"} err="failed to get container status \"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\": rpc error: code = NotFound desc = could not find container \"ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8\": container with ID starting with ec490f5a7cc631891fc13697c691fee753c916e379846f8d5d7f7e5176c8e0a8 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582623 4744 scope.go:117] "RemoveContainer" containerID="d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582836 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626"} err="failed to get container status \"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\": rpc error: code = NotFound desc = could not find container \"d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626\": container with ID starting with d5c1e94e95ab4d3ee6c85db254228ed678fa6adf1b7d72b7456f72dcd8af4626 not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.582858 4744 scope.go:117] "RemoveContainer" containerID="bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.583071 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff"} err="failed to get container status \"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\": rpc error: code = NotFound desc = could not find container \"bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff\": container with ID starting with bdc2d798594bffac7749e9dd8166e65bb77c2953aa997fd0e0f0a77f1b44a0ff not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.583097 4744 scope.go:117] "RemoveContainer" containerID="afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.583321 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f"} err="failed to get container status \"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\": rpc error: code = NotFound desc = could not find container \"afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f\": container with ID starting with afdc9a774fe5f6b9f39b210fd4719b5163fccef8e4e094552c6a26712be9bf2f not found: ID does not exist" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.583341 4744 scope.go:117] "RemoveContainer" containerID="5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6" Dec 01 08:28:25 crc kubenswrapper[4744]: I1201 08:28:25.583610 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6"} err="failed to get container status \"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\": rpc error: code = NotFound desc = could not find container \"5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6\": container with ID starting with 5b1af5ba12949c5d2fb7d8673e0b9b64eacd8db9f75493038758ebb52d2282e6 not found: ID does not exist" Dec 01 08:28:26 crc kubenswrapper[4744]: I1201 08:28:26.297735 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1df124eb-0511-4702-85d8-3e324c59bb0d" path="/var/lib/kubelet/pods/1df124eb-0511-4702-85d8-3e324c59bb0d/volumes" Dec 01 08:28:26 crc kubenswrapper[4744]: I1201 08:28:26.324037 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"0ddae094a7ca8e603b2d3c0b1112ecdcfd303389262e8f550721db5d83dbdf4f"} Dec 01 08:28:26 crc kubenswrapper[4744]: I1201 08:28:26.324334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"c96805b51195dbfe3313e2f52547cbe08a7ccf2d3a79e62f6a9f9956ffd2b46d"} Dec 01 08:28:26 crc kubenswrapper[4744]: I1201 08:28:26.324510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"2394692bdfa63b1dad9b56ae209864adeae612d34ac3d669f68480fa82b602a5"} Dec 01 08:28:26 crc kubenswrapper[4744]: I1201 08:28:26.324634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"e3fe5e5aa11e84371634edeb79312a6351af81ab921bd5933bdef814d68e5e7b"} Dec 01 08:28:26 crc kubenswrapper[4744]: I1201 08:28:26.324798 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"0d3022f717dd6a82d741ae1ac3568fadd63b169fe2fdbd361cceb89a3c0217ab"} Dec 01 08:28:26 crc kubenswrapper[4744]: I1201 08:28:26.324919 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"69016c1ac35908c1dc663295ffb0c56ff3043efdb522aa4e86b2d12ac0bc80b0"} Dec 01 08:28:29 crc kubenswrapper[4744]: I1201 08:28:29.353140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"f81b781dbe7ca060f5a095092ad98531e69d2e54cd83da1a73dc195bba5b377c"} Dec 01 08:28:31 crc kubenswrapper[4744]: I1201 08:28:31.371710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" event={"ID":"28e32de2-c0ce-426c-b30e-215073fba2bf","Type":"ContainerStarted","Data":"45c87b248e27be06088538206e80e064283e629421d8fafd1d4a1bcb656ec9fa"} Dec 01 08:28:31 crc kubenswrapper[4744]: I1201 08:28:31.372228 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:31 crc kubenswrapper[4744]: I1201 08:28:31.372241 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:31 crc kubenswrapper[4744]: I1201 08:28:31.372252 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:31 crc kubenswrapper[4744]: I1201 08:28:31.402035 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:31 crc kubenswrapper[4744]: I1201 08:28:31.402441 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:28:31 crc kubenswrapper[4744]: I1201 08:28:31.407784 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" podStartSLOduration=7.407766156 podStartE2EDuration="7.407766156s" podCreationTimestamp="2025-12-01 08:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:28:31.402062423 +0000 UTC m=+643.391120374" watchObservedRunningTime="2025-12-01 08:28:31.407766156 +0000 UTC m=+643.396824087" Dec 01 08:28:38 crc kubenswrapper[4744]: I1201 08:28:38.289603 4744 scope.go:117] "RemoveContainer" containerID="c5cd4b8744689d1838bd2d0e5d0166485469706e4e1b6d8f26489c876c06341e" Dec 01 08:28:38 crc kubenswrapper[4744]: E1201 08:28:38.290550 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vsh4r_openshift-multus(6a5e5532-8675-4f3f-aa30-7c22bc95b311)\"" pod="openshift-multus/multus-vsh4r" podUID="6a5e5532-8675-4f3f-aa30-7c22bc95b311" Dec 01 08:28:52 crc kubenswrapper[4744]: I1201 08:28:52.284539 4744 scope.go:117] "RemoveContainer" containerID="c5cd4b8744689d1838bd2d0e5d0166485469706e4e1b6d8f26489c876c06341e" Dec 01 08:28:52 crc kubenswrapper[4744]: I1201 08:28:52.516214 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vsh4r_6a5e5532-8675-4f3f-aa30-7c22bc95b311/kube-multus/2.log" Dec 01 08:28:52 crc kubenswrapper[4744]: I1201 08:28:52.516730 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vsh4r" event={"ID":"6a5e5532-8675-4f3f-aa30-7c22bc95b311","Type":"ContainerStarted","Data":"39e2ab47eb238bea7f45ade30ee2f2c0f8dd4fe2379211a53970902746d7723f"} Dec 01 08:28:54 crc kubenswrapper[4744]: I1201 08:28:54.829175 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hh8pr" Dec 01 08:29:00 crc kubenswrapper[4744]: I1201 08:29:00.969568 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2"] Dec 01 08:29:00 crc kubenswrapper[4744]: I1201 08:29:00.971040 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:00 crc kubenswrapper[4744]: I1201 08:29:00.973745 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 08:29:00 crc kubenswrapper[4744]: I1201 08:29:00.982062 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2"] Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.009656 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.010108 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv6j2\" (UniqueName: \"kubernetes.io/projected/34673b73-0a10-419d-8604-d8f595325c05-kube-api-access-zv6j2\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.010169 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.111395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv6j2\" (UniqueName: \"kubernetes.io/projected/34673b73-0a10-419d-8604-d8f595325c05-kube-api-access-zv6j2\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.111569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.111663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.112543 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.112880 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.142466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv6j2\" (UniqueName: \"kubernetes.io/projected/34673b73-0a10-419d-8604-d8f595325c05-kube-api-access-zv6j2\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.287988 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.494045 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2"] Dec 01 08:29:01 crc kubenswrapper[4744]: I1201 08:29:01.587756 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" event={"ID":"34673b73-0a10-419d-8604-d8f595325c05","Type":"ContainerStarted","Data":"d141ae6de959549a95ecd3283e26dfebb169379ce9ecc861cabbeb8f9c0f8e81"} Dec 01 08:29:02 crc kubenswrapper[4744]: I1201 08:29:02.594733 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" event={"ID":"34673b73-0a10-419d-8604-d8f595325c05","Type":"ContainerStarted","Data":"da08dacb770214642e3937a91e82b4f5ab232c1a0404664d8c8487d8313b0ff2"} Dec 01 08:29:03 crc kubenswrapper[4744]: I1201 08:29:03.600974 4744 generic.go:334] "Generic (PLEG): container finished" podID="34673b73-0a10-419d-8604-d8f595325c05" containerID="da08dacb770214642e3937a91e82b4f5ab232c1a0404664d8c8487d8313b0ff2" exitCode=0 Dec 01 08:29:03 crc kubenswrapper[4744]: I1201 08:29:03.601056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" event={"ID":"34673b73-0a10-419d-8604-d8f595325c05","Type":"ContainerDied","Data":"da08dacb770214642e3937a91e82b4f5ab232c1a0404664d8c8487d8313b0ff2"} Dec 01 08:29:04 crc kubenswrapper[4744]: I1201 08:29:04.614602 4744 generic.go:334] "Generic (PLEG): container finished" podID="34673b73-0a10-419d-8604-d8f595325c05" containerID="33b8afd97b3fff57eec9bc0fed07beef38c7f49c3e31fca2dc73da512e1a66b1" exitCode=0 Dec 01 08:29:04 crc kubenswrapper[4744]: I1201 08:29:04.614669 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" event={"ID":"34673b73-0a10-419d-8604-d8f595325c05","Type":"ContainerDied","Data":"33b8afd97b3fff57eec9bc0fed07beef38c7f49c3e31fca2dc73da512e1a66b1"} Dec 01 08:29:05 crc kubenswrapper[4744]: I1201 08:29:05.626588 4744 generic.go:334] "Generic (PLEG): container finished" podID="34673b73-0a10-419d-8604-d8f595325c05" containerID="3c1d0775a0ba93f9c63c01069ae84383d02c6804af21818b30312ccec7ab9b03" exitCode=0 Dec 01 08:29:05 crc kubenswrapper[4744]: I1201 08:29:05.626702 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" event={"ID":"34673b73-0a10-419d-8604-d8f595325c05","Type":"ContainerDied","Data":"3c1d0775a0ba93f9c63c01069ae84383d02c6804af21818b30312ccec7ab9b03"} Dec 01 08:29:06 crc kubenswrapper[4744]: I1201 08:29:06.955123 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.093431 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-bundle\") pod \"34673b73-0a10-419d-8604-d8f595325c05\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.093507 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv6j2\" (UniqueName: \"kubernetes.io/projected/34673b73-0a10-419d-8604-d8f595325c05-kube-api-access-zv6j2\") pod \"34673b73-0a10-419d-8604-d8f595325c05\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.093601 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-util\") pod \"34673b73-0a10-419d-8604-d8f595325c05\" (UID: \"34673b73-0a10-419d-8604-d8f595325c05\") " Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.095340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-bundle" (OuterVolumeSpecName: "bundle") pod "34673b73-0a10-419d-8604-d8f595325c05" (UID: "34673b73-0a10-419d-8604-d8f595325c05"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.102732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34673b73-0a10-419d-8604-d8f595325c05-kube-api-access-zv6j2" (OuterVolumeSpecName: "kube-api-access-zv6j2") pod "34673b73-0a10-419d-8604-d8f595325c05" (UID: "34673b73-0a10-419d-8604-d8f595325c05"). InnerVolumeSpecName "kube-api-access-zv6j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.128532 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-util" (OuterVolumeSpecName: "util") pod "34673b73-0a10-419d-8604-d8f595325c05" (UID: "34673b73-0a10-419d-8604-d8f595325c05"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.195718 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.195767 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34673b73-0a10-419d-8604-d8f595325c05-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.195785 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv6j2\" (UniqueName: \"kubernetes.io/projected/34673b73-0a10-419d-8604-d8f595325c05-kube-api-access-zv6j2\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.643267 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" event={"ID":"34673b73-0a10-419d-8604-d8f595325c05","Type":"ContainerDied","Data":"d141ae6de959549a95ecd3283e26dfebb169379ce9ecc861cabbeb8f9c0f8e81"} Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.643315 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d141ae6de959549a95ecd3283e26dfebb169379ce9ecc861cabbeb8f9c0f8e81" Dec 01 08:29:07 crc kubenswrapper[4744]: I1201 08:29:07.643443 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.671390 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq"] Dec 01 08:29:08 crc kubenswrapper[4744]: E1201 08:29:08.671889 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34673b73-0a10-419d-8604-d8f595325c05" containerName="extract" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.671901 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34673b73-0a10-419d-8604-d8f595325c05" containerName="extract" Dec 01 08:29:08 crc kubenswrapper[4744]: E1201 08:29:08.671918 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34673b73-0a10-419d-8604-d8f595325c05" containerName="util" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.671924 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34673b73-0a10-419d-8604-d8f595325c05" containerName="util" Dec 01 08:29:08 crc kubenswrapper[4744]: E1201 08:29:08.671936 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34673b73-0a10-419d-8604-d8f595325c05" containerName="pull" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.671942 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34673b73-0a10-419d-8604-d8f595325c05" containerName="pull" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.672046 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="34673b73-0a10-419d-8604-d8f595325c05" containerName="extract" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.672421 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.674167 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.674267 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-fv245" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.674855 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.687649 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq"] Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.828873 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skv2h\" (UniqueName: \"kubernetes.io/projected/ce67e6e1-9203-47bd-abb6-1dfe9b552bca-kube-api-access-skv2h\") pod \"nmstate-operator-5b5b58f5c8-mmbdq\" (UID: \"ce67e6e1-9203-47bd-abb6-1dfe9b552bca\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.930779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skv2h\" (UniqueName: \"kubernetes.io/projected/ce67e6e1-9203-47bd-abb6-1dfe9b552bca-kube-api-access-skv2h\") pod \"nmstate-operator-5b5b58f5c8-mmbdq\" (UID: \"ce67e6e1-9203-47bd-abb6-1dfe9b552bca\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" Dec 01 08:29:08 crc kubenswrapper[4744]: I1201 08:29:08.957631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skv2h\" (UniqueName: \"kubernetes.io/projected/ce67e6e1-9203-47bd-abb6-1dfe9b552bca-kube-api-access-skv2h\") pod \"nmstate-operator-5b5b58f5c8-mmbdq\" (UID: \"ce67e6e1-9203-47bd-abb6-1dfe9b552bca\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" Dec 01 08:29:09 crc kubenswrapper[4744]: I1201 08:29:09.027165 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" Dec 01 08:29:09 crc kubenswrapper[4744]: I1201 08:29:09.299347 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq"] Dec 01 08:29:09 crc kubenswrapper[4744]: W1201 08:29:09.315289 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce67e6e1_9203_47bd_abb6_1dfe9b552bca.slice/crio-afb75cf890ab59315cbe4ef8bf7ac96892cab9c9bab42b216f5237bea9d2eba5 WatchSource:0}: Error finding container afb75cf890ab59315cbe4ef8bf7ac96892cab9c9bab42b216f5237bea9d2eba5: Status 404 returned error can't find the container with id afb75cf890ab59315cbe4ef8bf7ac96892cab9c9bab42b216f5237bea9d2eba5 Dec 01 08:29:09 crc kubenswrapper[4744]: I1201 08:29:09.654812 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" event={"ID":"ce67e6e1-9203-47bd-abb6-1dfe9b552bca","Type":"ContainerStarted","Data":"afb75cf890ab59315cbe4ef8bf7ac96892cab9c9bab42b216f5237bea9d2eba5"} Dec 01 08:29:12 crc kubenswrapper[4744]: I1201 08:29:12.674519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" event={"ID":"ce67e6e1-9203-47bd-abb6-1dfe9b552bca","Type":"ContainerStarted","Data":"516b57afa653decb610362a60dc9993fd344e54dd6e4762983ddaca605625caf"} Dec 01 08:29:12 crc kubenswrapper[4744]: I1201 08:29:12.717856 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mmbdq" podStartSLOduration=2.842588654 podStartE2EDuration="4.717833472s" podCreationTimestamp="2025-12-01 08:29:08 +0000 UTC" firstStartedPulling="2025-12-01 08:29:09.31952774 +0000 UTC m=+681.308585661" lastFinishedPulling="2025-12-01 08:29:11.194772558 +0000 UTC m=+683.183830479" observedRunningTime="2025-12-01 08:29:12.696878017 +0000 UTC m=+684.685935938" watchObservedRunningTime="2025-12-01 08:29:12.717833472 +0000 UTC m=+684.706891393" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.761934 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lm574"] Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.763089 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.775645 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c"] Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.776664 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.777909 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-m8grd" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.783145 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.807164 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-7r6x7"] Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.807943 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.824426 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c"] Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.842045 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lm574"] Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.897792 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v"] Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.898925 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.904657 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54jsm\" (UniqueName: \"kubernetes.io/projected/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-kube-api-access-54jsm\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.904737 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkqc2\" (UniqueName: \"kubernetes.io/projected/5c7e45c3-5458-46ec-8c35-8e9776f6550e-kube-api-access-dkqc2\") pod \"nmstate-webhook-5f6d4c5ccb-gq68c\" (UID: \"5c7e45c3-5458-46ec-8c35-8e9776f6550e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.904800 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5c7e45c3-5458-46ec-8c35-8e9776f6550e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-gq68c\" (UID: \"5c7e45c3-5458-46ec-8c35-8e9776f6550e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.904902 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-dbus-socket\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.904955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9d70fe1c-9023-42af-9b1e-c078b39399f6-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.904990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d70fe1c-9023-42af-9b1e-c078b39399f6-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.905018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljks2\" (UniqueName: \"kubernetes.io/projected/9d70fe1c-9023-42af-9b1e-c078b39399f6-kube-api-access-ljks2\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.905045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-ovs-socket\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.905071 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-nmstate-lock\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.905190 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4gdl\" (UniqueName: \"kubernetes.io/projected/5cdd7453-8251-4776-82bf-b931ba787d45-kube-api-access-c4gdl\") pod \"nmstate-metrics-7f946cbc9-lm574\" (UID: \"5cdd7453-8251-4776-82bf-b931ba787d45\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.905917 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.905977 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-z578f" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.905995 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 01 08:29:13 crc kubenswrapper[4744]: I1201 08:29:13.908914 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v"] Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.006753 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4gdl\" (UniqueName: \"kubernetes.io/projected/5cdd7453-8251-4776-82bf-b931ba787d45-kube-api-access-c4gdl\") pod \"nmstate-metrics-7f946cbc9-lm574\" (UID: \"5cdd7453-8251-4776-82bf-b931ba787d45\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.006837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54jsm\" (UniqueName: \"kubernetes.io/projected/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-kube-api-access-54jsm\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.006889 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkqc2\" (UniqueName: \"kubernetes.io/projected/5c7e45c3-5458-46ec-8c35-8e9776f6550e-kube-api-access-dkqc2\") pod \"nmstate-webhook-5f6d4c5ccb-gq68c\" (UID: \"5c7e45c3-5458-46ec-8c35-8e9776f6550e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.006937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5c7e45c3-5458-46ec-8c35-8e9776f6550e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-gq68c\" (UID: \"5c7e45c3-5458-46ec-8c35-8e9776f6550e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.006984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-dbus-socket\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007015 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9d70fe1c-9023-42af-9b1e-c078b39399f6-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007042 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljks2\" (UniqueName: \"kubernetes.io/projected/9d70fe1c-9023-42af-9b1e-c078b39399f6-kube-api-access-ljks2\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007068 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d70fe1c-9023-42af-9b1e-c078b39399f6-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007099 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-ovs-socket\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-nmstate-lock\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007242 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-nmstate-lock\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-ovs-socket\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.007831 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-dbus-socket\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.008793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9d70fe1c-9023-42af-9b1e-c078b39399f6-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.023222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5c7e45c3-5458-46ec-8c35-8e9776f6550e-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-gq68c\" (UID: \"5c7e45c3-5458-46ec-8c35-8e9776f6550e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.023472 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d70fe1c-9023-42af-9b1e-c078b39399f6-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.046209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54jsm\" (UniqueName: \"kubernetes.io/projected/ea5a52f6-b418-4027-b6c0-3d57394d5fa5-kube-api-access-54jsm\") pod \"nmstate-handler-7r6x7\" (UID: \"ea5a52f6-b418-4027-b6c0-3d57394d5fa5\") " pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.053313 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4gdl\" (UniqueName: \"kubernetes.io/projected/5cdd7453-8251-4776-82bf-b931ba787d45-kube-api-access-c4gdl\") pod \"nmstate-metrics-7f946cbc9-lm574\" (UID: \"5cdd7453-8251-4776-82bf-b931ba787d45\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.053583 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkqc2\" (UniqueName: \"kubernetes.io/projected/5c7e45c3-5458-46ec-8c35-8e9776f6550e-kube-api-access-dkqc2\") pod \"nmstate-webhook-5f6d4c5ccb-gq68c\" (UID: \"5c7e45c3-5458-46ec-8c35-8e9776f6550e\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.055742 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljks2\" (UniqueName: \"kubernetes.io/projected/9d70fe1c-9023-42af-9b1e-c078b39399f6-kube-api-access-ljks2\") pod \"nmstate-console-plugin-7fbb5f6569-vrj4v\" (UID: \"9d70fe1c-9023-42af-9b1e-c078b39399f6\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.080859 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.098766 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-795d57c76f-rd8wf"] Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.105345 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.105837 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.115289 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-795d57c76f-rd8wf"] Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.129476 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.209925 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4r4x\" (UniqueName: \"kubernetes.io/projected/71b7feb7-843f-4926-8e9b-df472f6e033f-kube-api-access-c4r4x\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.209967 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-console-config\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.209986 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71b7feb7-843f-4926-8e9b-df472f6e033f-console-oauth-config\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.210003 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-oauth-serving-cert\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.210029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71b7feb7-843f-4926-8e9b-df472f6e033f-console-serving-cert\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.210060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-service-ca\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.210078 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-trusted-ca-bundle\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.218017 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.310603 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4r4x\" (UniqueName: \"kubernetes.io/projected/71b7feb7-843f-4926-8e9b-df472f6e033f-kube-api-access-c4r4x\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.310898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-console-config\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.310931 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71b7feb7-843f-4926-8e9b-df472f6e033f-console-oauth-config\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.310951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-oauth-serving-cert\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.310977 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71b7feb7-843f-4926-8e9b-df472f6e033f-console-serving-cert\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.311011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-service-ca\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.311030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-trusted-ca-bundle\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.311899 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-console-config\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.311903 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-oauth-serving-cert\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.312315 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-service-ca\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.313115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71b7feb7-843f-4926-8e9b-df472f6e033f-trusted-ca-bundle\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.317044 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71b7feb7-843f-4926-8e9b-df472f6e033f-console-serving-cert\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.318028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71b7feb7-843f-4926-8e9b-df472f6e033f-console-oauth-config\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.319994 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lm574"] Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.335963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4r4x\" (UniqueName: \"kubernetes.io/projected/71b7feb7-843f-4926-8e9b-df472f6e033f-kube-api-access-c4r4x\") pod \"console-795d57c76f-rd8wf\" (UID: \"71b7feb7-843f-4926-8e9b-df472f6e033f\") " pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.352183 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c"] Dec 01 08:29:14 crc kubenswrapper[4744]: W1201 08:29:14.354519 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c7e45c3_5458_46ec_8c35_8e9776f6550e.slice/crio-56669804ba1ee5ef6c5c717dfb7f15b40f095b15dd47b7c8ef7156e845780c8b WatchSource:0}: Error finding container 56669804ba1ee5ef6c5c717dfb7f15b40f095b15dd47b7c8ef7156e845780c8b: Status 404 returned error can't find the container with id 56669804ba1ee5ef6c5c717dfb7f15b40f095b15dd47b7c8ef7156e845780c8b Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.403950 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v"] Dec 01 08:29:14 crc kubenswrapper[4744]: W1201 08:29:14.407960 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d70fe1c_9023_42af_9b1e_c078b39399f6.slice/crio-a787ea36e3edff14f15e9ff20f0f9b606baab49c87265c788de25b7f77ea18d5 WatchSource:0}: Error finding container a787ea36e3edff14f15e9ff20f0f9b606baab49c87265c788de25b7f77ea18d5: Status 404 returned error can't find the container with id a787ea36e3edff14f15e9ff20f0f9b606baab49c87265c788de25b7f77ea18d5 Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.444726 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.686794 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" event={"ID":"5c7e45c3-5458-46ec-8c35-8e9776f6550e","Type":"ContainerStarted","Data":"56669804ba1ee5ef6c5c717dfb7f15b40f095b15dd47b7c8ef7156e845780c8b"} Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.688185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" event={"ID":"5cdd7453-8251-4776-82bf-b931ba787d45","Type":"ContainerStarted","Data":"d2e747597bac956814b4c66c92f95b2d2265bec4a96d2db72c61b4d6a6bc5166"} Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.689243 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" event={"ID":"9d70fe1c-9023-42af-9b1e-c078b39399f6","Type":"ContainerStarted","Data":"a787ea36e3edff14f15e9ff20f0f9b606baab49c87265c788de25b7f77ea18d5"} Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.690852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7r6x7" event={"ID":"ea5a52f6-b418-4027-b6c0-3d57394d5fa5","Type":"ContainerStarted","Data":"73b74a3ff501c083e05eaca67ec2c063e2ed216ff401c3a961a290f3c258ad28"} Dec 01 08:29:14 crc kubenswrapper[4744]: I1201 08:29:14.879638 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-795d57c76f-rd8wf"] Dec 01 08:29:14 crc kubenswrapper[4744]: W1201 08:29:14.887600 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71b7feb7_843f_4926_8e9b_df472f6e033f.slice/crio-230aa5c49e0ad845ad55d1baf47249b4de6b2d5eda0b3bd0bbfee8f047250abf WatchSource:0}: Error finding container 230aa5c49e0ad845ad55d1baf47249b4de6b2d5eda0b3bd0bbfee8f047250abf: Status 404 returned error can't find the container with id 230aa5c49e0ad845ad55d1baf47249b4de6b2d5eda0b3bd0bbfee8f047250abf Dec 01 08:29:15 crc kubenswrapper[4744]: I1201 08:29:15.700661 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-795d57c76f-rd8wf" event={"ID":"71b7feb7-843f-4926-8e9b-df472f6e033f","Type":"ContainerStarted","Data":"4846ac69b0d205def4a781622edcc320bee6904ff81418705c2061d7127a91be"} Dec 01 08:29:15 crc kubenswrapper[4744]: I1201 08:29:15.701209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-795d57c76f-rd8wf" event={"ID":"71b7feb7-843f-4926-8e9b-df472f6e033f","Type":"ContainerStarted","Data":"230aa5c49e0ad845ad55d1baf47249b4de6b2d5eda0b3bd0bbfee8f047250abf"} Dec 01 08:29:15 crc kubenswrapper[4744]: I1201 08:29:15.723160 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-795d57c76f-rd8wf" podStartSLOduration=1.723115581 podStartE2EDuration="1.723115581s" podCreationTimestamp="2025-12-01 08:29:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:29:15.722175384 +0000 UTC m=+687.711233305" watchObservedRunningTime="2025-12-01 08:29:15.723115581 +0000 UTC m=+687.712173502" Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.715610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" event={"ID":"5cdd7453-8251-4776-82bf-b931ba787d45","Type":"ContainerStarted","Data":"8f92685a9c57a231ca3be762cec1ba3a4fb106916f60c41b6fc7f53ea89dfd4a"} Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.717667 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" event={"ID":"9d70fe1c-9023-42af-9b1e-c078b39399f6","Type":"ContainerStarted","Data":"fd956399dd261e2f55a434e46a591f3e9e4b0a44c9f48a27ca56252e358fa87b"} Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.720495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7r6x7" event={"ID":"ea5a52f6-b418-4027-b6c0-3d57394d5fa5","Type":"ContainerStarted","Data":"9badd7330d12532f3dd76a65608d9b174f5ab806847d81d2fe9b7f01307e5bd9"} Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.720613 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.722148 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" event={"ID":"5c7e45c3-5458-46ec-8c35-8e9776f6550e","Type":"ContainerStarted","Data":"760ddedf50da08c9b2dd5a8d3e7c90efecf8e5c562bec3cc82d8e14eae0de581"} Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.722278 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.743525 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-vrj4v" podStartSLOduration=2.098660082 podStartE2EDuration="4.743508397s" podCreationTimestamp="2025-12-01 08:29:13 +0000 UTC" firstStartedPulling="2025-12-01 08:29:14.410210709 +0000 UTC m=+686.399268630" lastFinishedPulling="2025-12-01 08:29:17.055059024 +0000 UTC m=+689.044116945" observedRunningTime="2025-12-01 08:29:17.738291622 +0000 UTC m=+689.727349543" watchObservedRunningTime="2025-12-01 08:29:17.743508397 +0000 UTC m=+689.732566338" Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.758924 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" podStartSLOduration=2.029976617 podStartE2EDuration="4.758907477s" podCreationTimestamp="2025-12-01 08:29:13 +0000 UTC" firstStartedPulling="2025-12-01 08:29:14.358292661 +0000 UTC m=+686.347350582" lastFinishedPulling="2025-12-01 08:29:17.087223521 +0000 UTC m=+689.076281442" observedRunningTime="2025-12-01 08:29:17.758636789 +0000 UTC m=+689.747694760" watchObservedRunningTime="2025-12-01 08:29:17.758907477 +0000 UTC m=+689.747965398" Dec 01 08:29:17 crc kubenswrapper[4744]: I1201 08:29:17.785888 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-7r6x7" podStartSLOduration=1.9065285539999999 podStartE2EDuration="4.785822308s" podCreationTimestamp="2025-12-01 08:29:13 +0000 UTC" firstStartedPulling="2025-12-01 08:29:14.174274908 +0000 UTC m=+686.163332839" lastFinishedPulling="2025-12-01 08:29:17.053568632 +0000 UTC m=+689.042626593" observedRunningTime="2025-12-01 08:29:17.776132427 +0000 UTC m=+689.765190418" watchObservedRunningTime="2025-12-01 08:29:17.785822308 +0000 UTC m=+689.774880269" Dec 01 08:29:19 crc kubenswrapper[4744]: I1201 08:29:19.749839 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" event={"ID":"5cdd7453-8251-4776-82bf-b931ba787d45","Type":"ContainerStarted","Data":"7fb84d3d48a83f211cb3b59fdaba73907febea0f0d745e59f4ad779792cec429"} Dec 01 08:29:19 crc kubenswrapper[4744]: I1201 08:29:19.770075 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lm574" podStartSLOduration=1.873238385 podStartE2EDuration="6.770046285s" podCreationTimestamp="2025-12-01 08:29:13 +0000 UTC" firstStartedPulling="2025-12-01 08:29:14.330688761 +0000 UTC m=+686.319746682" lastFinishedPulling="2025-12-01 08:29:19.227496631 +0000 UTC m=+691.216554582" observedRunningTime="2025-12-01 08:29:19.76557702 +0000 UTC m=+691.754634991" watchObservedRunningTime="2025-12-01 08:29:19.770046285 +0000 UTC m=+691.759104256" Dec 01 08:29:24 crc kubenswrapper[4744]: I1201 08:29:24.166227 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-7r6x7" Dec 01 08:29:24 crc kubenswrapper[4744]: I1201 08:29:24.445203 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:24 crc kubenswrapper[4744]: I1201 08:29:24.445268 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:24 crc kubenswrapper[4744]: I1201 08:29:24.453597 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:24 crc kubenswrapper[4744]: I1201 08:29:24.793973 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-795d57c76f-rd8wf" Dec 01 08:29:24 crc kubenswrapper[4744]: I1201 08:29:24.875328 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nmgd9"] Dec 01 08:29:34 crc kubenswrapper[4744]: I1201 08:29:34.113188 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-gq68c" Dec 01 08:29:47 crc kubenswrapper[4744]: E1201 08:29:47.911968 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.140538 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb"] Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.141906 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.146220 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb"] Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.180892 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.282711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrg6v\" (UniqueName: \"kubernetes.io/projected/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-kube-api-access-zrg6v\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.282784 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.283052 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.384528 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.385131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrg6v\" (UniqueName: \"kubernetes.io/projected/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-kube-api-access-zrg6v\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.385249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.385901 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.385944 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.407365 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrg6v\" (UniqueName: \"kubernetes.io/projected/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-kube-api-access-zrg6v\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.508374 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.516274 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.562253 4744 scope.go:117] "RemoveContainer" containerID="882618f84cf5a287981fbf6984d8baed0720154c2aef509511915689166d78f6" Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.934809 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb"] Dec 01 08:29:48 crc kubenswrapper[4744]: I1201 08:29:48.951862 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" event={"ID":"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4","Type":"ContainerStarted","Data":"2d97f884830fdcbb6a232a9d3bdfe5cc155b9fae6a26d8f87814622110f68dbe"} Dec 01 08:29:49 crc kubenswrapper[4744]: I1201 08:29:49.934785 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-nmgd9" podUID="3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" containerName="console" containerID="cri-o://c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f" gracePeriod=15 Dec 01 08:29:49 crc kubenswrapper[4744]: I1201 08:29:49.960309 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerID="773d8cbc2ac11730e58a796251d9fc98df7728acd5c43ed84d2e786cfcfff5e0" exitCode=0 Dec 01 08:29:49 crc kubenswrapper[4744]: I1201 08:29:49.960376 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" event={"ID":"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4","Type":"ContainerDied","Data":"773d8cbc2ac11730e58a796251d9fc98df7728acd5c43ed84d2e786cfcfff5e0"} Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.355448 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nmgd9_3aa6496c-cce3-43bf-a8ec-a9b7a7891e10/console/0.log" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.355792 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.515905 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n65ff\" (UniqueName: \"kubernetes.io/projected/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-kube-api-access-n65ff\") pod \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.515982 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-serving-cert\") pod \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.516020 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-oauth-config\") pod \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.516251 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-service-ca\") pod \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.516304 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-config\") pod \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.516380 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-trusted-ca-bundle\") pod \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.516429 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-oauth-serving-cert\") pod \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\" (UID: \"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10\") " Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" (UID: "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517347 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-service-ca" (OuterVolumeSpecName: "service-ca") pod "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" (UID: "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517391 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" (UID: "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517468 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-config" (OuterVolumeSpecName: "console-config") pod "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" (UID: "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517745 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517774 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517794 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.517810 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.522472 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" (UID: "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.526067 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" (UID: "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.526125 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-kube-api-access-n65ff" (OuterVolumeSpecName: "kube-api-access-n65ff") pod "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" (UID: "3aa6496c-cce3-43bf-a8ec-a9b7a7891e10"). InnerVolumeSpecName "kube-api-access-n65ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.619384 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n65ff\" (UniqueName: \"kubernetes.io/projected/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-kube-api-access-n65ff\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.619771 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.619907 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.971608 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nmgd9_3aa6496c-cce3-43bf-a8ec-a9b7a7891e10/console/0.log" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.971684 4744 generic.go:334] "Generic (PLEG): container finished" podID="3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" containerID="c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f" exitCode=2 Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.971731 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nmgd9" event={"ID":"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10","Type":"ContainerDied","Data":"c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f"} Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.971772 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nmgd9" event={"ID":"3aa6496c-cce3-43bf-a8ec-a9b7a7891e10","Type":"ContainerDied","Data":"5dd76aafec99c08411ad2d2643b8b5bb842f980f8b571709e8702333e4b267c2"} Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.971800 4744 scope.go:117] "RemoveContainer" containerID="c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f" Dec 01 08:29:50 crc kubenswrapper[4744]: I1201 08:29:50.971971 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nmgd9" Dec 01 08:29:51 crc kubenswrapper[4744]: I1201 08:29:51.002008 4744 scope.go:117] "RemoveContainer" containerID="c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f" Dec 01 08:29:51 crc kubenswrapper[4744]: E1201 08:29:51.002611 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f\": container with ID starting with c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f not found: ID does not exist" containerID="c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f" Dec 01 08:29:51 crc kubenswrapper[4744]: I1201 08:29:51.002694 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f"} err="failed to get container status \"c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f\": rpc error: code = NotFound desc = could not find container \"c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f\": container with ID starting with c03e899a36ec2ef008de504fbc4433d97159bc6200b28a8b1ac882614931d28f not found: ID does not exist" Dec 01 08:29:51 crc kubenswrapper[4744]: I1201 08:29:51.012331 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nmgd9"] Dec 01 08:29:51 crc kubenswrapper[4744]: I1201 08:29:51.017607 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-nmgd9"] Dec 01 08:29:52 crc kubenswrapper[4744]: I1201 08:29:52.297844 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" path="/var/lib/kubelet/pods/3aa6496c-cce3-43bf-a8ec-a9b7a7891e10/volumes" Dec 01 08:29:52 crc kubenswrapper[4744]: I1201 08:29:52.991508 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" event={"ID":"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4","Type":"ContainerStarted","Data":"09820ab3e28446e7963e59abe57aada06664a83112700aa0c099939abd4f0de2"} Dec 01 08:29:54 crc kubenswrapper[4744]: I1201 08:29:54.000866 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerID="09820ab3e28446e7963e59abe57aada06664a83112700aa0c099939abd4f0de2" exitCode=0 Dec 01 08:29:54 crc kubenswrapper[4744]: I1201 08:29:54.000913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" event={"ID":"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4","Type":"ContainerDied","Data":"09820ab3e28446e7963e59abe57aada06664a83112700aa0c099939abd4f0de2"} Dec 01 08:29:55 crc kubenswrapper[4744]: I1201 08:29:55.017373 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerID="bcc1e28b809d61e470b21a08cc40e744b072b83855f105ed7d4977509ee03afe" exitCode=0 Dec 01 08:29:55 crc kubenswrapper[4744]: I1201 08:29:55.017462 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" event={"ID":"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4","Type":"ContainerDied","Data":"bcc1e28b809d61e470b21a08cc40e744b072b83855f105ed7d4977509ee03afe"} Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.350862 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.497493 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrg6v\" (UniqueName: \"kubernetes.io/projected/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-kube-api-access-zrg6v\") pod \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.497627 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-util\") pod \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.497762 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-bundle\") pod \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\" (UID: \"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4\") " Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.498882 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-bundle" (OuterVolumeSpecName: "bundle") pod "d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" (UID: "d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.503512 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-kube-api-access-zrg6v" (OuterVolumeSpecName: "kube-api-access-zrg6v") pod "d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" (UID: "d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4"). InnerVolumeSpecName "kube-api-access-zrg6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.512562 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-util" (OuterVolumeSpecName: "util") pod "d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" (UID: "d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.599751 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.599983 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:56 crc kubenswrapper[4744]: I1201 08:29:56.599997 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrg6v\" (UniqueName: \"kubernetes.io/projected/d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4-kube-api-access-zrg6v\") on node \"crc\" DevicePath \"\"" Dec 01 08:29:57 crc kubenswrapper[4744]: I1201 08:29:57.032703 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" event={"ID":"d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4","Type":"ContainerDied","Data":"2d97f884830fdcbb6a232a9d3bdfe5cc155b9fae6a26d8f87814622110f68dbe"} Dec 01 08:29:57 crc kubenswrapper[4744]: I1201 08:29:57.032763 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb" Dec 01 08:29:57 crc kubenswrapper[4744]: I1201 08:29:57.032772 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d97f884830fdcbb6a232a9d3bdfe5cc155b9fae6a26d8f87814622110f68dbe" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.142668 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt"] Dec 01 08:30:00 crc kubenswrapper[4744]: E1201 08:30:00.143331 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerName="pull" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.143353 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerName="pull" Dec 01 08:30:00 crc kubenswrapper[4744]: E1201 08:30:00.143400 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" containerName="console" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.143433 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" containerName="console" Dec 01 08:30:00 crc kubenswrapper[4744]: E1201 08:30:00.143452 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerName="util" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.143465 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerName="util" Dec 01 08:30:00 crc kubenswrapper[4744]: E1201 08:30:00.143489 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerName="extract" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.143500 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerName="extract" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.143649 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4" containerName="extract" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.143680 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa6496c-cce3-43bf-a8ec-a9b7a7891e10" containerName="console" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.144269 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.147370 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.148150 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.148748 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt"] Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.255166 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/832ffd1d-24fd-4170-b513-b9381be00bce-config-volume\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.255240 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxtxr\" (UniqueName: \"kubernetes.io/projected/832ffd1d-24fd-4170-b513-b9381be00bce-kube-api-access-fxtxr\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.255317 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/832ffd1d-24fd-4170-b513-b9381be00bce-secret-volume\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.356512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/832ffd1d-24fd-4170-b513-b9381be00bce-secret-volume\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.356593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/832ffd1d-24fd-4170-b513-b9381be00bce-config-volume\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.356636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxtxr\" (UniqueName: \"kubernetes.io/projected/832ffd1d-24fd-4170-b513-b9381be00bce-kube-api-access-fxtxr\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.358827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/832ffd1d-24fd-4170-b513-b9381be00bce-config-volume\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.362395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/832ffd1d-24fd-4170-b513-b9381be00bce-secret-volume\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.375336 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxtxr\" (UniqueName: \"kubernetes.io/projected/832ffd1d-24fd-4170-b513-b9381be00bce-kube-api-access-fxtxr\") pod \"collect-profiles-29409630-mklbt\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.462524 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:00 crc kubenswrapper[4744]: I1201 08:30:00.685115 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt"] Dec 01 08:30:01 crc kubenswrapper[4744]: I1201 08:30:01.053895 4744 generic.go:334] "Generic (PLEG): container finished" podID="832ffd1d-24fd-4170-b513-b9381be00bce" containerID="baea8f9cd6953899c97587599aab7427eb4d2c779e5ce2cfdac40736d46a606a" exitCode=0 Dec 01 08:30:01 crc kubenswrapper[4744]: I1201 08:30:01.054052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" event={"ID":"832ffd1d-24fd-4170-b513-b9381be00bce","Type":"ContainerDied","Data":"baea8f9cd6953899c97587599aab7427eb4d2c779e5ce2cfdac40736d46a606a"} Dec 01 08:30:01 crc kubenswrapper[4744]: I1201 08:30:01.054209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" event={"ID":"832ffd1d-24fd-4170-b513-b9381be00bce","Type":"ContainerStarted","Data":"97942fbddc3790fa633dccaa515785d7b25efce1466e45139ad5ad956c7062ec"} Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.365703 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.483236 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/832ffd1d-24fd-4170-b513-b9381be00bce-secret-volume\") pod \"832ffd1d-24fd-4170-b513-b9381be00bce\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.483346 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxtxr\" (UniqueName: \"kubernetes.io/projected/832ffd1d-24fd-4170-b513-b9381be00bce-kube-api-access-fxtxr\") pod \"832ffd1d-24fd-4170-b513-b9381be00bce\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.483529 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/832ffd1d-24fd-4170-b513-b9381be00bce-config-volume\") pod \"832ffd1d-24fd-4170-b513-b9381be00bce\" (UID: \"832ffd1d-24fd-4170-b513-b9381be00bce\") " Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.484623 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832ffd1d-24fd-4170-b513-b9381be00bce-config-volume" (OuterVolumeSpecName: "config-volume") pod "832ffd1d-24fd-4170-b513-b9381be00bce" (UID: "832ffd1d-24fd-4170-b513-b9381be00bce"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.487840 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/832ffd1d-24fd-4170-b513-b9381be00bce-kube-api-access-fxtxr" (OuterVolumeSpecName: "kube-api-access-fxtxr") pod "832ffd1d-24fd-4170-b513-b9381be00bce" (UID: "832ffd1d-24fd-4170-b513-b9381be00bce"). InnerVolumeSpecName "kube-api-access-fxtxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.488145 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832ffd1d-24fd-4170-b513-b9381be00bce-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "832ffd1d-24fd-4170-b513-b9381be00bce" (UID: "832ffd1d-24fd-4170-b513-b9381be00bce"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.585388 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/832ffd1d-24fd-4170-b513-b9381be00bce-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.585460 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/832ffd1d-24fd-4170-b513-b9381be00bce-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:30:02 crc kubenswrapper[4744]: I1201 08:30:02.585478 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxtxr\" (UniqueName: \"kubernetes.io/projected/832ffd1d-24fd-4170-b513-b9381be00bce-kube-api-access-fxtxr\") on node \"crc\" DevicePath \"\"" Dec 01 08:30:03 crc kubenswrapper[4744]: I1201 08:30:03.064927 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" event={"ID":"832ffd1d-24fd-4170-b513-b9381be00bce","Type":"ContainerDied","Data":"97942fbddc3790fa633dccaa515785d7b25efce1466e45139ad5ad956c7062ec"} Dec 01 08:30:03 crc kubenswrapper[4744]: I1201 08:30:03.064963 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97942fbddc3790fa633dccaa515785d7b25efce1466e45139ad5ad956c7062ec" Dec 01 08:30:03 crc kubenswrapper[4744]: I1201 08:30:03.064978 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.245254 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg"] Dec 01 08:30:06 crc kubenswrapper[4744]: E1201 08:30:06.245838 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832ffd1d-24fd-4170-b513-b9381be00bce" containerName="collect-profiles" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.245854 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="832ffd1d-24fd-4170-b513-b9381be00bce" containerName="collect-profiles" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.245983 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="832ffd1d-24fd-4170-b513-b9381be00bce" containerName="collect-profiles" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.246432 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.248346 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.248401 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.248672 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.248753 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.254401 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg"] Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.258547 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-wrgdz" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.333593 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-apiservice-cert\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.333664 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-webhook-cert\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.333682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2c94\" (UniqueName: \"kubernetes.io/projected/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-kube-api-access-r2c94\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.434536 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-apiservice-cert\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.434608 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-webhook-cert\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.434630 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2c94\" (UniqueName: \"kubernetes.io/projected/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-kube-api-access-r2c94\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.444145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-apiservice-cert\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.445895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-webhook-cert\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.454078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2c94\" (UniqueName: \"kubernetes.io/projected/6d5bd225-d8b4-4836-bb19-7e5c0fd540ae-kube-api-access-r2c94\") pod \"metallb-operator-controller-manager-79b5ddfc8f-x7dgg\" (UID: \"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae\") " pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.491748 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g"] Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.492821 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.494391 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.494710 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-tvlwp" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.495075 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.506192 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g"] Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.535585 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-webhook-cert\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.535678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rrh8\" (UniqueName: \"kubernetes.io/projected/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-kube-api-access-8rrh8\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.535723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-apiservice-cert\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.562338 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.636570 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-webhook-cert\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.636609 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rrh8\" (UniqueName: \"kubernetes.io/projected/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-kube-api-access-8rrh8\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.636656 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-apiservice-cert\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.642059 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-apiservice-cert\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.642625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-webhook-cert\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.653135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rrh8\" (UniqueName: \"kubernetes.io/projected/84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d-kube-api-access-8rrh8\") pod \"metallb-operator-webhook-server-68f89dd57b-kgd5g\" (UID: \"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d\") " pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.816382 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg"] Dec 01 08:30:06 crc kubenswrapper[4744]: I1201 08:30:06.816560 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:06 crc kubenswrapper[4744]: W1201 08:30:06.833795 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d5bd225_d8b4_4836_bb19_7e5c0fd540ae.slice/crio-b4bf449650365e8a582b14fe86cf1db221272b699e06e56d2b6afce259bf68dd WatchSource:0}: Error finding container b4bf449650365e8a582b14fe86cf1db221272b699e06e56d2b6afce259bf68dd: Status 404 returned error can't find the container with id b4bf449650365e8a582b14fe86cf1db221272b699e06e56d2b6afce259bf68dd Dec 01 08:30:07 crc kubenswrapper[4744]: I1201 08:30:07.088753 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" event={"ID":"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae","Type":"ContainerStarted","Data":"b4bf449650365e8a582b14fe86cf1db221272b699e06e56d2b6afce259bf68dd"} Dec 01 08:30:07 crc kubenswrapper[4744]: I1201 08:30:07.089373 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g"] Dec 01 08:30:07 crc kubenswrapper[4744]: W1201 08:30:07.098030 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d30a9f_3fb4_4a8c_bc3a_cf9a45720d1d.slice/crio-ce5b81e9a7cb4f8fbd57e3cdbe953a641e2ab2ae458fb73c08030ad5921a4d7a WatchSource:0}: Error finding container ce5b81e9a7cb4f8fbd57e3cdbe953a641e2ab2ae458fb73c08030ad5921a4d7a: Status 404 returned error can't find the container with id ce5b81e9a7cb4f8fbd57e3cdbe953a641e2ab2ae458fb73c08030ad5921a4d7a Dec 01 08:30:08 crc kubenswrapper[4744]: I1201 08:30:08.094345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" event={"ID":"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d","Type":"ContainerStarted","Data":"ce5b81e9a7cb4f8fbd57e3cdbe953a641e2ab2ae458fb73c08030ad5921a4d7a"} Dec 01 08:30:11 crc kubenswrapper[4744]: I1201 08:30:11.118888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" event={"ID":"6d5bd225-d8b4-4836-bb19-7e5c0fd540ae","Type":"ContainerStarted","Data":"1f2a83cafe07e94d8e76cb03cf6f8853aa7e7e86c83ab99108c909e42e1d612a"} Dec 01 08:30:11 crc kubenswrapper[4744]: I1201 08:30:11.119341 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:13 crc kubenswrapper[4744]: I1201 08:30:13.135930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" event={"ID":"84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d","Type":"ContainerStarted","Data":"2f4d97e2afd5d6509a76af5eda69e0ed1fcbea895ee6ff138a75294c2ddcdfa5"} Dec 01 08:30:13 crc kubenswrapper[4744]: I1201 08:30:13.136454 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:13 crc kubenswrapper[4744]: I1201 08:30:13.171068 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" podStartSLOduration=1.759175342 podStartE2EDuration="7.171036553s" podCreationTimestamp="2025-12-01 08:30:06 +0000 UTC" firstStartedPulling="2025-12-01 08:30:07.100755188 +0000 UTC m=+739.089813109" lastFinishedPulling="2025-12-01 08:30:12.512616389 +0000 UTC m=+744.501674320" observedRunningTime="2025-12-01 08:30:13.164129709 +0000 UTC m=+745.153187660" watchObservedRunningTime="2025-12-01 08:30:13.171036553 +0000 UTC m=+745.160094514" Dec 01 08:30:13 crc kubenswrapper[4744]: I1201 08:30:13.171968 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" podStartSLOduration=3.471945247 podStartE2EDuration="7.171957559s" podCreationTimestamp="2025-12-01 08:30:06 +0000 UTC" firstStartedPulling="2025-12-01 08:30:06.835667839 +0000 UTC m=+738.824725760" lastFinishedPulling="2025-12-01 08:30:10.535680151 +0000 UTC m=+742.524738072" observedRunningTime="2025-12-01 08:30:11.140578965 +0000 UTC m=+743.129636916" watchObservedRunningTime="2025-12-01 08:30:13.171957559 +0000 UTC m=+745.161015510" Dec 01 08:30:26 crc kubenswrapper[4744]: I1201 08:30:26.821401 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-68f89dd57b-kgd5g" Dec 01 08:30:28 crc kubenswrapper[4744]: I1201 08:30:28.335800 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:30:46 crc kubenswrapper[4744]: I1201 08:30:46.565239 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-79b5ddfc8f-x7dgg" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.291471 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wqpdx"] Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.293467 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.295624 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.295720 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.295880 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-hkmmr" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.295895 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq"] Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.297730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.299254 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.308654 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq"] Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.381603 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6vzqh"] Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.382681 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.386322 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.386322 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dstl6" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.386497 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.386838 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.388772 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-sockets\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.388866 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-metrics\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.388898 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-conf\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.388936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc5np\" (UniqueName: \"kubernetes.io/projected/33198d7f-46c9-41db-a7ce-4916d55561f2-kube-api-access-lc5np\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.389128 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-reloader\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.389163 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33198d7f-46c9-41db-a7ce-4916d55561f2-metrics-certs\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.389188 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppl2q\" (UniqueName: \"kubernetes.io/projected/481b9c20-e688-4def-a63b-57038302c23c-kube-api-access-ppl2q\") pod \"frr-k8s-webhook-server-7fcb986d4-c6jzq\" (UID: \"481b9c20-e688-4def-a63b-57038302c23c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.389229 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/481b9c20-e688-4def-a63b-57038302c23c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c6jzq\" (UID: \"481b9c20-e688-4def-a63b-57038302c23c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.389314 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-startup\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.402230 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-bwdwl"] Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.403345 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.407937 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.418832 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-bwdwl"] Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490822 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-metrics\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-conf\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490886 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc5np\" (UniqueName: \"kubernetes.io/projected/33198d7f-46c9-41db-a7ce-4916d55561f2-kube-api-access-lc5np\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-metrics-certs\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490933 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-reloader\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33198d7f-46c9-41db-a7ce-4916d55561f2-metrics-certs\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490981 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppl2q\" (UniqueName: \"kubernetes.io/projected/481b9c20-e688-4def-a63b-57038302c23c-kube-api-access-ppl2q\") pod \"frr-k8s-webhook-server-7fcb986d4-c6jzq\" (UID: \"481b9c20-e688-4def-a63b-57038302c23c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.490996 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-cert\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491012 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/481b9c20-e688-4def-a63b-57038302c23c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c6jzq\" (UID: \"481b9c20-e688-4def-a63b-57038302c23c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5xdn\" (UniqueName: \"kubernetes.io/projected/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-kube-api-access-w5xdn\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491056 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngqss\" (UniqueName: \"kubernetes.io/projected/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-kube-api-access-ngqss\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491076 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-startup\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491091 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metrics-certs\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491107 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metallb-excludel2\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491148 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-sockets\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491211 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-metrics\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491465 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-sockets\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-conf\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.491725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33198d7f-46c9-41db-a7ce-4916d55561f2-reloader\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.492336 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33198d7f-46c9-41db-a7ce-4916d55561f2-frr-startup\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.500340 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33198d7f-46c9-41db-a7ce-4916d55561f2-metrics-certs\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.505852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/481b9c20-e688-4def-a63b-57038302c23c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c6jzq\" (UID: \"481b9c20-e688-4def-a63b-57038302c23c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.508368 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppl2q\" (UniqueName: \"kubernetes.io/projected/481b9c20-e688-4def-a63b-57038302c23c-kube-api-access-ppl2q\") pod \"frr-k8s-webhook-server-7fcb986d4-c6jzq\" (UID: \"481b9c20-e688-4def-a63b-57038302c23c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.511870 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc5np\" (UniqueName: \"kubernetes.io/projected/33198d7f-46c9-41db-a7ce-4916d55561f2-kube-api-access-lc5np\") pod \"frr-k8s-wqpdx\" (UID: \"33198d7f-46c9-41db-a7ce-4916d55561f2\") " pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.592397 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-metrics-certs\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.592484 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-cert\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.592519 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5xdn\" (UniqueName: \"kubernetes.io/projected/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-kube-api-access-w5xdn\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.592552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngqss\" (UniqueName: \"kubernetes.io/projected/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-kube-api-access-ngqss\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.592580 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.592606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metrics-certs\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.592631 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metallb-excludel2\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: E1201 08:30:47.593281 4744 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 01 08:30:47 crc kubenswrapper[4744]: E1201 08:30:47.593345 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metrics-certs podName:3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c nodeName:}" failed. No retries permitted until 2025-12-01 08:30:48.093326629 +0000 UTC m=+780.082384550 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metrics-certs") pod "speaker-6vzqh" (UID: "3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c") : secret "speaker-certs-secret" not found Dec 01 08:30:47 crc kubenswrapper[4744]: E1201 08:30:47.593347 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 08:30:47 crc kubenswrapper[4744]: E1201 08:30:47.593469 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist podName:3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c nodeName:}" failed. No retries permitted until 2025-12-01 08:30:48.093442172 +0000 UTC m=+780.082500173 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist") pod "speaker-6vzqh" (UID: "3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c") : secret "metallb-memberlist" not found Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.593502 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metallb-excludel2\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.596886 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-cert\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.597001 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-metrics-certs\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.609053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngqss\" (UniqueName: \"kubernetes.io/projected/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-kube-api-access-ngqss\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.618867 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.627721 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.628567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5xdn\" (UniqueName: \"kubernetes.io/projected/edc692c2-6d9e-42fd-b54a-f856dd9f2ed1-kube-api-access-w5xdn\") pod \"controller-f8648f98b-bwdwl\" (UID: \"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1\") " pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.721319 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:47 crc kubenswrapper[4744]: I1201 08:30:47.931400 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq"] Dec 01 08:30:47 crc kubenswrapper[4744]: W1201 08:30:47.943318 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod481b9c20_e688_4def_a63b_57038302c23c.slice/crio-1a62fd711a8aa24fd75b9cfec9668961eef558cab721344a68df1fbd8bffad8f WatchSource:0}: Error finding container 1a62fd711a8aa24fd75b9cfec9668961eef558cab721344a68df1fbd8bffad8f: Status 404 returned error can't find the container with id 1a62fd711a8aa24fd75b9cfec9668961eef558cab721344a68df1fbd8bffad8f Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.100511 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metrics-certs\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.100565 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:48 crc kubenswrapper[4744]: E1201 08:30:48.100798 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 08:30:48 crc kubenswrapper[4744]: E1201 08:30:48.100870 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist podName:3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c nodeName:}" failed. No retries permitted until 2025-12-01 08:30:49.100850433 +0000 UTC m=+781.089908354 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist") pod "speaker-6vzqh" (UID: "3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c") : secret "metallb-memberlist" not found Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.107127 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-metrics-certs\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.204920 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-bwdwl"] Dec 01 08:30:48 crc kubenswrapper[4744]: W1201 08:30:48.211165 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedc692c2_6d9e_42fd_b54a_f856dd9f2ed1.slice/crio-767f50766387fa3171bb4d3aa36804820b187a2df3aefd65579544524bfa842e WatchSource:0}: Error finding container 767f50766387fa3171bb4d3aa36804820b187a2df3aefd65579544524bfa842e: Status 404 returned error can't find the container with id 767f50766387fa3171bb4d3aa36804820b187a2df3aefd65579544524bfa842e Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.360784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" event={"ID":"481b9c20-e688-4def-a63b-57038302c23c","Type":"ContainerStarted","Data":"1a62fd711a8aa24fd75b9cfec9668961eef558cab721344a68df1fbd8bffad8f"} Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.362532 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerStarted","Data":"57f3be33a1c4da61e2913f7ed46d933e1cf9560e5d5affdb5a7f6029c3a9b12a"} Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.363899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bwdwl" event={"ID":"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1","Type":"ContainerStarted","Data":"767f50766387fa3171bb4d3aa36804820b187a2df3aefd65579544524bfa842e"} Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.634148 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:30:48 crc kubenswrapper[4744]: I1201 08:30:48.634210 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.122340 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.129245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c-memberlist\") pod \"speaker-6vzqh\" (UID: \"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c\") " pod="metallb-system/speaker-6vzqh" Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.198542 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dstl6" Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.206509 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6vzqh" Dec 01 08:30:49 crc kubenswrapper[4744]: W1201 08:30:49.232971 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3be29ff8_9ae8_41be_bd0b_9a0ba9ce747c.slice/crio-40f1322119309b01f95e888945a4a79a4a455abfbd081c10ce9a984ec6f0fc60 WatchSource:0}: Error finding container 40f1322119309b01f95e888945a4a79a4a455abfbd081c10ce9a984ec6f0fc60: Status 404 returned error can't find the container with id 40f1322119309b01f95e888945a4a79a4a455abfbd081c10ce9a984ec6f0fc60 Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.375093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bwdwl" event={"ID":"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1","Type":"ContainerStarted","Data":"868d0d5fb8e299b0ce9624057aff514a011157963d0ae5bd0b30d4995e79aa53"} Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.375157 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bwdwl" event={"ID":"edc692c2-6d9e-42fd-b54a-f856dd9f2ed1","Type":"ContainerStarted","Data":"0139b9bef9850d03ede0db1e6a2749f5cb4cde07884fc37d21dd617e0c229f20"} Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.375219 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.379135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6vzqh" event={"ID":"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c","Type":"ContainerStarted","Data":"40f1322119309b01f95e888945a4a79a4a455abfbd081c10ce9a984ec6f0fc60"} Dec 01 08:30:49 crc kubenswrapper[4744]: I1201 08:30:49.412696 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-bwdwl" podStartSLOduration=2.412673256 podStartE2EDuration="2.412673256s" podCreationTimestamp="2025-12-01 08:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:30:49.397249725 +0000 UTC m=+781.386307636" watchObservedRunningTime="2025-12-01 08:30:49.412673256 +0000 UTC m=+781.401731177" Dec 01 08:30:50 crc kubenswrapper[4744]: I1201 08:30:50.393623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6vzqh" event={"ID":"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c","Type":"ContainerStarted","Data":"414469968d5844a1382ddfa241c53512c44b781512730473d466fdba516f90eb"} Dec 01 08:30:50 crc kubenswrapper[4744]: I1201 08:30:50.393926 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6vzqh" event={"ID":"3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c","Type":"ContainerStarted","Data":"fff78b2fa9bf6ba2c2be7e8c2caf030d57922870eb2a3174e8cb98aff34a86c7"} Dec 01 08:30:50 crc kubenswrapper[4744]: I1201 08:30:50.393972 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6vzqh" Dec 01 08:30:50 crc kubenswrapper[4744]: I1201 08:30:50.415609 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6vzqh" podStartSLOduration=3.4155908999999998 podStartE2EDuration="3.4155909s" podCreationTimestamp="2025-12-01 08:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:30:50.411550037 +0000 UTC m=+782.400607968" watchObservedRunningTime="2025-12-01 08:30:50.4155909 +0000 UTC m=+782.404648811" Dec 01 08:30:55 crc kubenswrapper[4744]: I1201 08:30:55.443771 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" event={"ID":"481b9c20-e688-4def-a63b-57038302c23c","Type":"ContainerStarted","Data":"d074b27a88fb81c96a7ddb06a45c0c4dac649abf162da433ec3ce3694260448d"} Dec 01 08:30:55 crc kubenswrapper[4744]: I1201 08:30:55.444435 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:30:55 crc kubenswrapper[4744]: I1201 08:30:55.446252 4744 generic.go:334] "Generic (PLEG): container finished" podID="33198d7f-46c9-41db-a7ce-4916d55561f2" containerID="fc1439a2144ce970d8879bb04e77851fbeeaddab97c46dafc868d2d18db8c914" exitCode=0 Dec 01 08:30:55 crc kubenswrapper[4744]: I1201 08:30:55.446324 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerDied","Data":"fc1439a2144ce970d8879bb04e77851fbeeaddab97c46dafc868d2d18db8c914"} Dec 01 08:30:55 crc kubenswrapper[4744]: I1201 08:30:55.504067 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" podStartSLOduration=1.601158929 podStartE2EDuration="8.504038935s" podCreationTimestamp="2025-12-01 08:30:47 +0000 UTC" firstStartedPulling="2025-12-01 08:30:47.946544238 +0000 UTC m=+779.935602159" lastFinishedPulling="2025-12-01 08:30:54.849424244 +0000 UTC m=+786.838482165" observedRunningTime="2025-12-01 08:30:55.467631058 +0000 UTC m=+787.456688999" watchObservedRunningTime="2025-12-01 08:30:55.504038935 +0000 UTC m=+787.493096896" Dec 01 08:30:56 crc kubenswrapper[4744]: I1201 08:30:56.459223 4744 generic.go:334] "Generic (PLEG): container finished" podID="33198d7f-46c9-41db-a7ce-4916d55561f2" containerID="d28838f456c26e02618e8cfd501e15845842bd47f0e35cca78b167b990fa3416" exitCode=0 Dec 01 08:30:56 crc kubenswrapper[4744]: I1201 08:30:56.459325 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerDied","Data":"d28838f456c26e02618e8cfd501e15845842bd47f0e35cca78b167b990fa3416"} Dec 01 08:30:57 crc kubenswrapper[4744]: I1201 08:30:57.465810 4744 generic.go:334] "Generic (PLEG): container finished" podID="33198d7f-46c9-41db-a7ce-4916d55561f2" containerID="021647520b99343fa02614a5b61528bbd4d20d8e8e66242a31b2fc3b51d55e56" exitCode=0 Dec 01 08:30:57 crc kubenswrapper[4744]: I1201 08:30:57.465866 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerDied","Data":"021647520b99343fa02614a5b61528bbd4d20d8e8e66242a31b2fc3b51d55e56"} Dec 01 08:30:58 crc kubenswrapper[4744]: I1201 08:30:58.475529 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerStarted","Data":"1b7716385abd12f42483494f0d7993ce181ef4475394dbe427dd8e0f3f505984"} Dec 01 08:30:58 crc kubenswrapper[4744]: I1201 08:30:58.475938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerStarted","Data":"bd313dca1160a822507c020514c7cb2e15622db37f92be4a85df4936162ffc11"} Dec 01 08:30:58 crc kubenswrapper[4744]: I1201 08:30:58.475949 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerStarted","Data":"cfaffacdf5bb6f41b73e6c1b0e89ba58e5f5d6fb6eb1a0ee851c814465d43789"} Dec 01 08:30:58 crc kubenswrapper[4744]: I1201 08:30:58.475960 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerStarted","Data":"b846846ffb5a110f751350f6fe19692879e5e167376c8c1fd22cd9eb4a604c8a"} Dec 01 08:30:58 crc kubenswrapper[4744]: I1201 08:30:58.475971 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerStarted","Data":"e086a31a7cad68d7c018656ce25771af1aec5f41352311855b45855cb542fb67"} Dec 01 08:30:59 crc kubenswrapper[4744]: I1201 08:30:59.210427 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6vzqh" Dec 01 08:30:59 crc kubenswrapper[4744]: I1201 08:30:59.492463 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wqpdx" event={"ID":"33198d7f-46c9-41db-a7ce-4916d55561f2","Type":"ContainerStarted","Data":"9048f795225b8a565e9be66674283baddf429f5ac33a957590799ab561ce8afe"} Dec 01 08:30:59 crc kubenswrapper[4744]: I1201 08:30:59.492830 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.261237 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wqpdx" podStartSLOduration=8.257327364 podStartE2EDuration="15.26121861s" podCreationTimestamp="2025-12-01 08:30:47 +0000 UTC" firstStartedPulling="2025-12-01 08:30:47.864375355 +0000 UTC m=+779.853433276" lastFinishedPulling="2025-12-01 08:30:54.868266601 +0000 UTC m=+786.857324522" observedRunningTime="2025-12-01 08:30:59.526947957 +0000 UTC m=+791.516005888" watchObservedRunningTime="2025-12-01 08:31:02.26121861 +0000 UTC m=+794.250276531" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.264509 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-jjn9d"] Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.265226 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjn9d" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.270024 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.270671 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.272310 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-nzxsk" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.282169 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jjn9d"] Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.418896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgjz6\" (UniqueName: \"kubernetes.io/projected/c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad-kube-api-access-qgjz6\") pod \"openstack-operator-index-jjn9d\" (UID: \"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad\") " pod="openstack-operators/openstack-operator-index-jjn9d" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.520508 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgjz6\" (UniqueName: \"kubernetes.io/projected/c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad-kube-api-access-qgjz6\") pod \"openstack-operator-index-jjn9d\" (UID: \"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad\") " pod="openstack-operators/openstack-operator-index-jjn9d" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.537939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgjz6\" (UniqueName: \"kubernetes.io/projected/c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad-kube-api-access-qgjz6\") pod \"openstack-operator-index-jjn9d\" (UID: \"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad\") " pod="openstack-operators/openstack-operator-index-jjn9d" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.606456 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjn9d" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.619912 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.669056 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:31:02 crc kubenswrapper[4744]: I1201 08:31:02.819182 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jjn9d"] Dec 01 08:31:02 crc kubenswrapper[4744]: W1201 08:31:02.829158 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2a0f5c6_6aa7_449b_aa88_3b4df470b7ad.slice/crio-4c706a81688d17b05d057fdd7fdb2f96b8a8bda2d3fd65963e4aece582a20608 WatchSource:0}: Error finding container 4c706a81688d17b05d057fdd7fdb2f96b8a8bda2d3fd65963e4aece582a20608: Status 404 returned error can't find the container with id 4c706a81688d17b05d057fdd7fdb2f96b8a8bda2d3fd65963e4aece582a20608 Dec 01 08:31:03 crc kubenswrapper[4744]: I1201 08:31:03.530334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjn9d" event={"ID":"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad","Type":"ContainerStarted","Data":"4c706a81688d17b05d057fdd7fdb2f96b8a8bda2d3fd65963e4aece582a20608"} Dec 01 08:31:05 crc kubenswrapper[4744]: I1201 08:31:05.639505 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-jjn9d"] Dec 01 08:31:06 crc kubenswrapper[4744]: I1201 08:31:06.247035 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-q498w"] Dec 01 08:31:06 crc kubenswrapper[4744]: I1201 08:31:06.250514 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:06 crc kubenswrapper[4744]: I1201 08:31:06.260752 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q498w"] Dec 01 08:31:06 crc kubenswrapper[4744]: I1201 08:31:06.378299 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94qwr\" (UniqueName: \"kubernetes.io/projected/34cb91f3-fee7-4916-be15-b257cab23850-kube-api-access-94qwr\") pod \"openstack-operator-index-q498w\" (UID: \"34cb91f3-fee7-4916-be15-b257cab23850\") " pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:06 crc kubenswrapper[4744]: I1201 08:31:06.480611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94qwr\" (UniqueName: \"kubernetes.io/projected/34cb91f3-fee7-4916-be15-b257cab23850-kube-api-access-94qwr\") pod \"openstack-operator-index-q498w\" (UID: \"34cb91f3-fee7-4916-be15-b257cab23850\") " pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:06 crc kubenswrapper[4744]: I1201 08:31:06.529178 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94qwr\" (UniqueName: \"kubernetes.io/projected/34cb91f3-fee7-4916-be15-b257cab23850-kube-api-access-94qwr\") pod \"openstack-operator-index-q498w\" (UID: \"34cb91f3-fee7-4916-be15-b257cab23850\") " pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:06 crc kubenswrapper[4744]: I1201 08:31:06.585924 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.314032 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q498w"] Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.560030 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q498w" event={"ID":"34cb91f3-fee7-4916-be15-b257cab23850","Type":"ContainerStarted","Data":"74eda4c710bcce0ca00c7c977ed1400e9d998e255d61850d580da4a0baa5fa29"} Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.560094 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q498w" event={"ID":"34cb91f3-fee7-4916-be15-b257cab23850","Type":"ContainerStarted","Data":"e4cc09a1e9b2f6c91e3271685fe29a334be03cab8bd3d6fec17e6241435b3df1"} Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.563004 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjn9d" event={"ID":"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad","Type":"ContainerStarted","Data":"c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1"} Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.563015 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-jjn9d" podUID="c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad" containerName="registry-server" containerID="cri-o://c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1" gracePeriod=2 Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.583722 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-q498w" podStartSLOduration=1.540213259 podStartE2EDuration="1.583702945s" podCreationTimestamp="2025-12-01 08:31:06 +0000 UTC" firstStartedPulling="2025-12-01 08:31:07.323496624 +0000 UTC m=+799.312554585" lastFinishedPulling="2025-12-01 08:31:07.36698634 +0000 UTC m=+799.356044271" observedRunningTime="2025-12-01 08:31:07.581337529 +0000 UTC m=+799.570395450" watchObservedRunningTime="2025-12-01 08:31:07.583702945 +0000 UTC m=+799.572760876" Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.604294 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-jjn9d" podStartSLOduration=1.495798498 podStartE2EDuration="5.60427073s" podCreationTimestamp="2025-12-01 08:31:02 +0000 UTC" firstStartedPulling="2025-12-01 08:31:02.831197117 +0000 UTC m=+794.820255038" lastFinishedPulling="2025-12-01 08:31:06.939669349 +0000 UTC m=+798.928727270" observedRunningTime="2025-12-01 08:31:07.599399674 +0000 UTC m=+799.588457605" watchObservedRunningTime="2025-12-01 08:31:07.60427073 +0000 UTC m=+799.593328651" Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.622628 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wqpdx" Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.633814 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c6jzq" Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.725018 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-bwdwl" Dec 01 08:31:07 crc kubenswrapper[4744]: I1201 08:31:07.932452 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjn9d" Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.109594 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgjz6\" (UniqueName: \"kubernetes.io/projected/c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad-kube-api-access-qgjz6\") pod \"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad\" (UID: \"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad\") " Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.115845 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad-kube-api-access-qgjz6" (OuterVolumeSpecName: "kube-api-access-qgjz6") pod "c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad" (UID: "c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad"). InnerVolumeSpecName "kube-api-access-qgjz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.211643 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgjz6\" (UniqueName: \"kubernetes.io/projected/c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad-kube-api-access-qgjz6\") on node \"crc\" DevicePath \"\"" Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.572254 4744 generic.go:334] "Generic (PLEG): container finished" podID="c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad" containerID="c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1" exitCode=0 Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.572349 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjn9d" Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.572357 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjn9d" event={"ID":"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad","Type":"ContainerDied","Data":"c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1"} Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.572482 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjn9d" event={"ID":"c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad","Type":"ContainerDied","Data":"4c706a81688d17b05d057fdd7fdb2f96b8a8bda2d3fd65963e4aece582a20608"} Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.572526 4744 scope.go:117] "RemoveContainer" containerID="c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1" Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.603515 4744 scope.go:117] "RemoveContainer" containerID="c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1" Dec 01 08:31:08 crc kubenswrapper[4744]: E1201 08:31:08.604023 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1\": container with ID starting with c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1 not found: ID does not exist" containerID="c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1" Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.604064 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1"} err="failed to get container status \"c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1\": rpc error: code = NotFound desc = could not find container \"c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1\": container with ID starting with c340b06840e2af283e7c3ef305e9de3e454987cd282d4339e6f74b7bb840aac1 not found: ID does not exist" Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.605930 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-jjn9d"] Dec 01 08:31:08 crc kubenswrapper[4744]: I1201 08:31:08.610298 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-jjn9d"] Dec 01 08:31:10 crc kubenswrapper[4744]: I1201 08:31:10.307486 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad" path="/var/lib/kubelet/pods/c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad/volumes" Dec 01 08:31:16 crc kubenswrapper[4744]: I1201 08:31:16.586712 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:16 crc kubenswrapper[4744]: I1201 08:31:16.587611 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:16 crc kubenswrapper[4744]: I1201 08:31:16.624009 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:16 crc kubenswrapper[4744]: I1201 08:31:16.671879 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-q498w" Dec 01 08:31:17 crc kubenswrapper[4744]: I1201 08:31:17.903459 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p"] Dec 01 08:31:17 crc kubenswrapper[4744]: E1201 08:31:17.903791 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad" containerName="registry-server" Dec 01 08:31:17 crc kubenswrapper[4744]: I1201 08:31:17.903812 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad" containerName="registry-server" Dec 01 08:31:17 crc kubenswrapper[4744]: I1201 08:31:17.904041 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2a0f5c6-6aa7-449b-aa88-3b4df470b7ad" containerName="registry-server" Dec 01 08:31:17 crc kubenswrapper[4744]: I1201 08:31:17.905493 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:17 crc kubenswrapper[4744]: I1201 08:31:17.907856 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-585xg" Dec 01 08:31:17 crc kubenswrapper[4744]: I1201 08:31:17.921067 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p"] Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.047848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wvnz\" (UniqueName: \"kubernetes.io/projected/d667dc57-45ab-4792-ac76-f7de322dbb88-kube-api-access-7wvnz\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.047968 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-util\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.048045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-bundle\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.149049 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-util\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.149178 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-bundle\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.149265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wvnz\" (UniqueName: \"kubernetes.io/projected/d667dc57-45ab-4792-ac76-f7de322dbb88-kube-api-access-7wvnz\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.149916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-bundle\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.150036 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-util\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.182941 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wvnz\" (UniqueName: \"kubernetes.io/projected/d667dc57-45ab-4792-ac76-f7de322dbb88-kube-api-access-7wvnz\") pod \"9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.230101 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.634523 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.634841 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:31:18 crc kubenswrapper[4744]: I1201 08:31:18.694620 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p"] Dec 01 08:31:18 crc kubenswrapper[4744]: W1201 08:31:18.704010 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd667dc57_45ab_4792_ac76_f7de322dbb88.slice/crio-3f980db909d2731ea4fd88b1589a053933c111d747e47bcee2d08b9ef2830d63 WatchSource:0}: Error finding container 3f980db909d2731ea4fd88b1589a053933c111d747e47bcee2d08b9ef2830d63: Status 404 returned error can't find the container with id 3f980db909d2731ea4fd88b1589a053933c111d747e47bcee2d08b9ef2830d63 Dec 01 08:31:19 crc kubenswrapper[4744]: I1201 08:31:19.665094 4744 generic.go:334] "Generic (PLEG): container finished" podID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerID="dac43a72393dfb219510d75ad90e1aab3d5006cda5c149b107d49202a65d498b" exitCode=0 Dec 01 08:31:19 crc kubenswrapper[4744]: I1201 08:31:19.665481 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" event={"ID":"d667dc57-45ab-4792-ac76-f7de322dbb88","Type":"ContainerDied","Data":"dac43a72393dfb219510d75ad90e1aab3d5006cda5c149b107d49202a65d498b"} Dec 01 08:31:19 crc kubenswrapper[4744]: I1201 08:31:19.665522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" event={"ID":"d667dc57-45ab-4792-ac76-f7de322dbb88","Type":"ContainerStarted","Data":"3f980db909d2731ea4fd88b1589a053933c111d747e47bcee2d08b9ef2830d63"} Dec 01 08:31:20 crc kubenswrapper[4744]: I1201 08:31:20.676762 4744 generic.go:334] "Generic (PLEG): container finished" podID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerID="a67798c4aeadd98442086b31a2f33ec1a3460154c6589905272743910aecb498" exitCode=0 Dec 01 08:31:20 crc kubenswrapper[4744]: I1201 08:31:20.677235 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" event={"ID":"d667dc57-45ab-4792-ac76-f7de322dbb88","Type":"ContainerDied","Data":"a67798c4aeadd98442086b31a2f33ec1a3460154c6589905272743910aecb498"} Dec 01 08:31:21 crc kubenswrapper[4744]: I1201 08:31:21.694864 4744 generic.go:334] "Generic (PLEG): container finished" podID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerID="585a038a2e55ddedf80e57e328f7d2d644eedb231e8e7b9a2c9f4819e40d197d" exitCode=0 Dec 01 08:31:21 crc kubenswrapper[4744]: I1201 08:31:21.694928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" event={"ID":"d667dc57-45ab-4792-ac76-f7de322dbb88","Type":"ContainerDied","Data":"585a038a2e55ddedf80e57e328f7d2d644eedb231e8e7b9a2c9f4819e40d197d"} Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.001727 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.025515 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wvnz\" (UniqueName: \"kubernetes.io/projected/d667dc57-45ab-4792-ac76-f7de322dbb88-kube-api-access-7wvnz\") pod \"d667dc57-45ab-4792-ac76-f7de322dbb88\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.025664 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-bundle\") pod \"d667dc57-45ab-4792-ac76-f7de322dbb88\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.025723 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-util\") pod \"d667dc57-45ab-4792-ac76-f7de322dbb88\" (UID: \"d667dc57-45ab-4792-ac76-f7de322dbb88\") " Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.028010 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-bundle" (OuterVolumeSpecName: "bundle") pod "d667dc57-45ab-4792-ac76-f7de322dbb88" (UID: "d667dc57-45ab-4792-ac76-f7de322dbb88"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.041648 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d667dc57-45ab-4792-ac76-f7de322dbb88-kube-api-access-7wvnz" (OuterVolumeSpecName: "kube-api-access-7wvnz") pod "d667dc57-45ab-4792-ac76-f7de322dbb88" (UID: "d667dc57-45ab-4792-ac76-f7de322dbb88"). InnerVolumeSpecName "kube-api-access-7wvnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.044464 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-util" (OuterVolumeSpecName: "util") pod "d667dc57-45ab-4792-ac76-f7de322dbb88" (UID: "d667dc57-45ab-4792-ac76-f7de322dbb88"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.127248 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wvnz\" (UniqueName: \"kubernetes.io/projected/d667dc57-45ab-4792-ac76-f7de322dbb88-kube-api-access-7wvnz\") on node \"crc\" DevicePath \"\"" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.127280 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.127292 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d667dc57-45ab-4792-ac76-f7de322dbb88-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.716604 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" event={"ID":"d667dc57-45ab-4792-ac76-f7de322dbb88","Type":"ContainerDied","Data":"3f980db909d2731ea4fd88b1589a053933c111d747e47bcee2d08b9ef2830d63"} Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.717026 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f980db909d2731ea4fd88b1589a053933c111d747e47bcee2d08b9ef2830d63" Dec 01 08:31:23 crc kubenswrapper[4744]: I1201 08:31:23.716676 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.267928 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt"] Dec 01 08:31:25 crc kubenswrapper[4744]: E1201 08:31:25.268130 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerName="extract" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.268140 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerName="extract" Dec 01 08:31:25 crc kubenswrapper[4744]: E1201 08:31:25.268160 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerName="pull" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.268166 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerName="pull" Dec 01 08:31:25 crc kubenswrapper[4744]: E1201 08:31:25.268179 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerName="util" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.268185 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerName="util" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.268282 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d667dc57-45ab-4792-ac76-f7de322dbb88" containerName="extract" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.268706 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.279065 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-d5mhs" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.289882 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt"] Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.357682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w28sj\" (UniqueName: \"kubernetes.io/projected/3df0fc73-2812-42d2-9c7f-e1f33591e8a1-kube-api-access-w28sj\") pod \"openstack-operator-controller-operator-7d866d68c9-2fxjt\" (UID: \"3df0fc73-2812-42d2-9c7f-e1f33591e8a1\") " pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.458913 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w28sj\" (UniqueName: \"kubernetes.io/projected/3df0fc73-2812-42d2-9c7f-e1f33591e8a1-kube-api-access-w28sj\") pod \"openstack-operator-controller-operator-7d866d68c9-2fxjt\" (UID: \"3df0fc73-2812-42d2-9c7f-e1f33591e8a1\") " pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.478165 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w28sj\" (UniqueName: \"kubernetes.io/projected/3df0fc73-2812-42d2-9c7f-e1f33591e8a1-kube-api-access-w28sj\") pod \"openstack-operator-controller-operator-7d866d68c9-2fxjt\" (UID: \"3df0fc73-2812-42d2-9c7f-e1f33591e8a1\") " pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" Dec 01 08:31:25 crc kubenswrapper[4744]: I1201 08:31:25.582888 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" Dec 01 08:31:26 crc kubenswrapper[4744]: I1201 08:31:26.008987 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt"] Dec 01 08:31:26 crc kubenswrapper[4744]: I1201 08:31:26.739208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" event={"ID":"3df0fc73-2812-42d2-9c7f-e1f33591e8a1","Type":"ContainerStarted","Data":"097aa0568d4103a3243643c8283f21b0a84dc711ad34ec818b247f1fa9291183"} Dec 01 08:31:29 crc kubenswrapper[4744]: I1201 08:31:29.769853 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" event={"ID":"3df0fc73-2812-42d2-9c7f-e1f33591e8a1","Type":"ContainerStarted","Data":"3b6f1ec670324a007859be16865a6266e7f895404c99192305ff5be29823900b"} Dec 01 08:31:29 crc kubenswrapper[4744]: I1201 08:31:29.770454 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" Dec 01 08:31:29 crc kubenswrapper[4744]: I1201 08:31:29.808500 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" podStartSLOduration=1.345078411 podStartE2EDuration="4.808478978s" podCreationTimestamp="2025-12-01 08:31:25 +0000 UTC" firstStartedPulling="2025-12-01 08:31:26.01939031 +0000 UTC m=+818.008448231" lastFinishedPulling="2025-12-01 08:31:29.482790867 +0000 UTC m=+821.471848798" observedRunningTime="2025-12-01 08:31:29.805288839 +0000 UTC m=+821.794346760" watchObservedRunningTime="2025-12-01 08:31:29.808478978 +0000 UTC m=+821.797536909" Dec 01 08:31:35 crc kubenswrapper[4744]: I1201 08:31:35.586270 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7d866d68c9-2fxjt" Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.636010 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.637329 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.637549 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.638474 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d6c923f3edaf0f41eb1f1e74b19db369be4219783160ca17cb5ceacdf850175"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.638753 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://0d6c923f3edaf0f41eb1f1e74b19db369be4219783160ca17cb5ceacdf850175" gracePeriod=600 Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.913148 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="0d6c923f3edaf0f41eb1f1e74b19db369be4219783160ca17cb5ceacdf850175" exitCode=0 Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.913238 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"0d6c923f3edaf0f41eb1f1e74b19db369be4219783160ca17cb5ceacdf850175"} Dec 01 08:31:48 crc kubenswrapper[4744]: I1201 08:31:48.913674 4744 scope.go:117] "RemoveContainer" containerID="08978ca5ab125cd2f75e7461dc8ec1a4a02dd4cdb1c39a55fa75057d29c93966" Dec 01 08:31:49 crc kubenswrapper[4744]: I1201 08:31:49.925398 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"64ec6b8815aa74952cf65128020a2c3112aeb4c1255a0d0fff478a4f4a871f94"} Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.794830 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.796165 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.800720 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qmxx5" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.813427 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.827591 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.828540 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.831080 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jmvxw" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.833763 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.837597 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.850558 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ds94h" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.863242 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.864467 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.864759 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.869192 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9hr6l" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.876309 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.877188 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.889002 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.892834 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-2tvkl" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.893827 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.898609 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.907347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzzhj\" (UniqueName: \"kubernetes.io/projected/7bc8fb90-3b43-43f8-8df4-128c639ad76a-kube-api-access-mzzhj\") pod \"barbican-operator-controller-manager-7d9dfd778-wf77m\" (UID: \"7bc8fb90-3b43-43f8-8df4-128c639ad76a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.907472 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.908335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.913872 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-nlzc5" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.922849 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.931559 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.932720 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.942029 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.942321 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-frt77" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.944489 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.945467 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.948733 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-r2dtz" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.952248 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.965933 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.976572 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v"] Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.977900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.987555 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mqjvl" Dec 01 08:31:54 crc kubenswrapper[4744]: I1201 08:31:54.993088 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.008670 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtxpz\" (UniqueName: \"kubernetes.io/projected/f5fa46a8-32d1-4eda-89ec-04bdfd35ea78-kube-api-access-rtxpz\") pod \"glance-operator-controller-manager-668d9c48b9-kvcrb\" (UID: \"f5fa46a8-32d1-4eda-89ec-04bdfd35ea78\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.008748 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47nmk\" (UniqueName: \"kubernetes.io/projected/badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a-kube-api-access-47nmk\") pod \"heat-operator-controller-manager-5f64f6f8bb-7z2ml\" (UID: \"badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.008913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6fcb\" (UniqueName: \"kubernetes.io/projected/d9f50601-c7cd-4db6-82dd-9581a5dc8a10-kube-api-access-w6fcb\") pod \"cinder-operator-controller-manager-859b6ccc6-lkltp\" (UID: \"d9f50601-c7cd-4db6-82dd-9581a5dc8a10\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.008981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcwqd\" (UniqueName: \"kubernetes.io/projected/2fbe6884-845e-4c7d-aff5-0deb483f4d19-kube-api-access-jcwqd\") pod \"designate-operator-controller-manager-78b4bc895b-8zfjd\" (UID: \"2fbe6884-845e-4c7d-aff5-0deb483f4d19\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.009138 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzzhj\" (UniqueName: \"kubernetes.io/projected/7bc8fb90-3b43-43f8-8df4-128c639ad76a-kube-api-access-mzzhj\") pod \"barbican-operator-controller-manager-7d9dfd778-wf77m\" (UID: \"7bc8fb90-3b43-43f8-8df4-128c639ad76a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.009183 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4n6z\" (UniqueName: \"kubernetes.io/projected/5f4c0ab8-4dee-4c39-b630-061580e95592-kube-api-access-g4n6z\") pod \"horizon-operator-controller-manager-68c6d99b8f-pv6f6\" (UID: \"5f4c0ab8-4dee-4c39-b630-061580e95592\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.017765 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.018801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.020886 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.021058 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-dtg57" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.088860 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110359 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b7rg\" (UniqueName: \"kubernetes.io/projected/4cf33df0-316f-4acf-b105-8dc967fae633-kube-api-access-6b7rg\") pod \"manila-operator-controller-manager-6546668bfd-tlszh\" (UID: \"4cf33df0-316f-4acf-b105-8dc967fae633\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110448 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4n6z\" (UniqueName: \"kubernetes.io/projected/5f4c0ab8-4dee-4c39-b630-061580e95592-kube-api-access-g4n6z\") pod \"horizon-operator-controller-manager-68c6d99b8f-pv6f6\" (UID: \"5f4c0ab8-4dee-4c39-b630-061580e95592\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110474 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtxpz\" (UniqueName: \"kubernetes.io/projected/f5fa46a8-32d1-4eda-89ec-04bdfd35ea78-kube-api-access-rtxpz\") pod \"glance-operator-controller-manager-668d9c48b9-kvcrb\" (UID: \"f5fa46a8-32d1-4eda-89ec-04bdfd35ea78\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110497 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7qk5\" (UniqueName: \"kubernetes.io/projected/8b89e5df-c38d-4465-aa74-b7c67e268a0d-kube-api-access-v7qk5\") pod \"keystone-operator-controller-manager-546d4bdf48-7fn8v\" (UID: \"8b89e5df-c38d-4465-aa74-b7c67e268a0d\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110533 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47nmk\" (UniqueName: \"kubernetes.io/projected/badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a-kube-api-access-47nmk\") pod \"heat-operator-controller-manager-5f64f6f8bb-7z2ml\" (UID: \"badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110559 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcwqd\" (UniqueName: \"kubernetes.io/projected/2fbe6884-845e-4c7d-aff5-0deb483f4d19-kube-api-access-jcwqd\") pod \"designate-operator-controller-manager-78b4bc895b-8zfjd\" (UID: \"2fbe6884-845e-4c7d-aff5-0deb483f4d19\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110578 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6fcb\" (UniqueName: \"kubernetes.io/projected/d9f50601-c7cd-4db6-82dd-9581a5dc8a10-kube-api-access-w6fcb\") pod \"cinder-operator-controller-manager-859b6ccc6-lkltp\" (UID: \"d9f50601-c7cd-4db6-82dd-9581a5dc8a10\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110600 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xxts\" (UniqueName: \"kubernetes.io/projected/214d7853-ed12-4624-82cd-6b279a827a13-kube-api-access-6xxts\") pod \"ironic-operator-controller-manager-6c548fd776-gg7s7\" (UID: \"214d7853-ed12-4624-82cd-6b279a827a13\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110631 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw85m\" (UniqueName: \"kubernetes.io/projected/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-kube-api-access-cw85m\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.110650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.111450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzzhj\" (UniqueName: \"kubernetes.io/projected/7bc8fb90-3b43-43f8-8df4-128c639ad76a-kube-api-access-mzzhj\") pod \"barbican-operator-controller-manager-7d9dfd778-wf77m\" (UID: \"7bc8fb90-3b43-43f8-8df4-128c639ad76a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.111624 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.115134 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-h9v4v" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.120208 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.130103 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.146297 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcwqd\" (UniqueName: \"kubernetes.io/projected/2fbe6884-845e-4c7d-aff5-0deb483f4d19-kube-api-access-jcwqd\") pod \"designate-operator-controller-manager-78b4bc895b-8zfjd\" (UID: \"2fbe6884-845e-4c7d-aff5-0deb483f4d19\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.148845 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.152770 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.156380 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtxpz\" (UniqueName: \"kubernetes.io/projected/f5fa46a8-32d1-4eda-89ec-04bdfd35ea78-kube-api-access-rtxpz\") pod \"glance-operator-controller-manager-668d9c48b9-kvcrb\" (UID: \"f5fa46a8-32d1-4eda-89ec-04bdfd35ea78\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.156543 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-54kpw" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.157321 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4n6z\" (UniqueName: \"kubernetes.io/projected/5f4c0ab8-4dee-4c39-b630-061580e95592-kube-api-access-g4n6z\") pod \"horizon-operator-controller-manager-68c6d99b8f-pv6f6\" (UID: \"5f4c0ab8-4dee-4c39-b630-061580e95592\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.159739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6fcb\" (UniqueName: \"kubernetes.io/projected/d9f50601-c7cd-4db6-82dd-9581a5dc8a10-kube-api-access-w6fcb\") pod \"cinder-operator-controller-manager-859b6ccc6-lkltp\" (UID: \"d9f50601-c7cd-4db6-82dd-9581a5dc8a10\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.164889 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.165946 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.167710 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-v6xpq" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.168971 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47nmk\" (UniqueName: \"kubernetes.io/projected/badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a-kube-api-access-47nmk\") pod \"heat-operator-controller-manager-5f64f6f8bb-7z2ml\" (UID: \"badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.176741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.181715 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.199168 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.202752 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.213299 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7qk5\" (UniqueName: \"kubernetes.io/projected/8b89e5df-c38d-4465-aa74-b7c67e268a0d-kube-api-access-v7qk5\") pod \"keystone-operator-controller-manager-546d4bdf48-7fn8v\" (UID: \"8b89e5df-c38d-4465-aa74-b7c67e268a0d\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.213440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp7bq\" (UniqueName: \"kubernetes.io/projected/f68be396-2f8f-4c90-99d3-bbe12b7d6c8e-kube-api-access-vp7bq\") pod \"mariadb-operator-controller-manager-56bbcc9d85-wqvcv\" (UID: \"f68be396-2f8f-4c90-99d3-bbe12b7d6c8e\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.213492 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xxts\" (UniqueName: \"kubernetes.io/projected/214d7853-ed12-4624-82cd-6b279a827a13-kube-api-access-6xxts\") pod \"ironic-operator-controller-manager-6c548fd776-gg7s7\" (UID: \"214d7853-ed12-4624-82cd-6b279a827a13\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.213526 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw85m\" (UniqueName: \"kubernetes.io/projected/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-kube-api-access-cw85m\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.213544 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.213561 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b7rg\" (UniqueName: \"kubernetes.io/projected/4cf33df0-316f-4acf-b105-8dc967fae633-kube-api-access-6b7rg\") pod \"manila-operator-controller-manager-6546668bfd-tlszh\" (UID: \"4cf33df0-316f-4acf-b105-8dc967fae633\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.214182 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.214217 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert podName:ceb61e3f-5534-46ad-ba39-1fe0b07530f7 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:55.714203266 +0000 UTC m=+847.703261177 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert") pod "infra-operator-controller-manager-57548d458d-dgxbb" (UID: "ceb61e3f-5534-46ad-ba39-1fe0b07530f7") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.214696 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-htgsw"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.215669 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.221775 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-xft6j" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.222501 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.236309 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.246478 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-htgsw"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.247855 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xxts\" (UniqueName: \"kubernetes.io/projected/214d7853-ed12-4624-82cd-6b279a827a13-kube-api-access-6xxts\") pod \"ironic-operator-controller-manager-6c548fd776-gg7s7\" (UID: \"214d7853-ed12-4624-82cd-6b279a827a13\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.250230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b7rg\" (UniqueName: \"kubernetes.io/projected/4cf33df0-316f-4acf-b105-8dc967fae633-kube-api-access-6b7rg\") pod \"manila-operator-controller-manager-6546668bfd-tlszh\" (UID: \"4cf33df0-316f-4acf-b105-8dc967fae633\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.251625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7qk5\" (UniqueName: \"kubernetes.io/projected/8b89e5df-c38d-4465-aa74-b7c67e268a0d-kube-api-access-v7qk5\") pod \"keystone-operator-controller-manager-546d4bdf48-7fn8v\" (UID: \"8b89e5df-c38d-4465-aa74-b7c67e268a0d\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.252117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw85m\" (UniqueName: \"kubernetes.io/projected/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-kube-api-access-cw85m\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.253719 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.267500 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.268873 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.277548 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.278274 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-bfqmh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.295374 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.315040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp7bq\" (UniqueName: \"kubernetes.io/projected/f68be396-2f8f-4c90-99d3-bbe12b7d6c8e-kube-api-access-vp7bq\") pod \"mariadb-operator-controller-manager-56bbcc9d85-wqvcv\" (UID: \"f68be396-2f8f-4c90-99d3-bbe12b7d6c8e\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.315085 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csl84\" (UniqueName: \"kubernetes.io/projected/5e8a8cd5-653f-4ef0-97c6-36b1a045f59a-kube-api-access-csl84\") pod \"nova-operator-controller-manager-697bc559fc-q9dxh\" (UID: \"5e8a8cd5-653f-4ef0-97c6-36b1a045f59a\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.315175 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r58g7\" (UniqueName: \"kubernetes.io/projected/b42166df-9137-4ddb-b9b0-33515fcab298-kube-api-access-r58g7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rbt96\" (UID: \"b42166df-9137-4ddb-b9b0-33515fcab298\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.315201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58v6c\" (UniqueName: \"kubernetes.io/projected/c0f26427-9c16-40b5-ad62-5a7161cba83c-kube-api-access-58v6c\") pod \"octavia-operator-controller-manager-998648c74-htgsw\" (UID: \"c0f26427-9c16-40b5-ad62-5a7161cba83c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.319504 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.320717 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.327485 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-848rm"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.328600 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.331581 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.334932 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-lvdvb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.335168 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2dsdb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.335550 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.343496 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-848rm"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.357714 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.364550 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.373085 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp7bq\" (UniqueName: \"kubernetes.io/projected/f68be396-2f8f-4c90-99d3-bbe12b7d6c8e-kube-api-access-vp7bq\") pod \"mariadb-operator-controller-manager-56bbcc9d85-wqvcv\" (UID: \"f68be396-2f8f-4c90-99d3-bbe12b7d6c8e\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.377334 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.378581 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.380573 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-2lbrl" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.413029 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.417947 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vnmk\" (UniqueName: \"kubernetes.io/projected/83c42810-1f8a-4616-9544-db48c6020c85-kube-api-access-2vnmk\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.418007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csl84\" (UniqueName: \"kubernetes.io/projected/5e8a8cd5-653f-4ef0-97c6-36b1a045f59a-kube-api-access-csl84\") pod \"nova-operator-controller-manager-697bc559fc-q9dxh\" (UID: \"5e8a8cd5-653f-4ef0-97c6-36b1a045f59a\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.418054 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfl46\" (UniqueName: \"kubernetes.io/projected/102b5fb4-119f-41ba-845e-1b691b08eb57-kube-api-access-mfl46\") pod \"ovn-operator-controller-manager-b6456fdb6-fjx5c\" (UID: \"102b5fb4-119f-41ba-845e-1b691b08eb57\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.418101 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q498d\" (UniqueName: \"kubernetes.io/projected/cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5-kube-api-access-q498d\") pod \"placement-operator-controller-manager-78f8948974-848rm\" (UID: \"cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.418124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.418172 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r58g7\" (UniqueName: \"kubernetes.io/projected/b42166df-9137-4ddb-b9b0-33515fcab298-kube-api-access-r58g7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rbt96\" (UID: \"b42166df-9137-4ddb-b9b0-33515fcab298\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.418202 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58v6c\" (UniqueName: \"kubernetes.io/projected/c0f26427-9c16-40b5-ad62-5a7161cba83c-kube-api-access-58v6c\") pod \"octavia-operator-controller-manager-998648c74-htgsw\" (UID: \"c0f26427-9c16-40b5-ad62-5a7161cba83c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.444628 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.452463 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r58g7\" (UniqueName: \"kubernetes.io/projected/b42166df-9137-4ddb-b9b0-33515fcab298-kube-api-access-r58g7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rbt96\" (UID: \"b42166df-9137-4ddb-b9b0-33515fcab298\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.452545 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csl84\" (UniqueName: \"kubernetes.io/projected/5e8a8cd5-653f-4ef0-97c6-36b1a045f59a-kube-api-access-csl84\") pod \"nova-operator-controller-manager-697bc559fc-q9dxh\" (UID: \"5e8a8cd5-653f-4ef0-97c6-36b1a045f59a\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.455735 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58v6c\" (UniqueName: \"kubernetes.io/projected/c0f26427-9c16-40b5-ad62-5a7161cba83c-kube-api-access-58v6c\") pod \"octavia-operator-controller-manager-998648c74-htgsw\" (UID: \"c0f26427-9c16-40b5-ad62-5a7161cba83c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.459716 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.460668 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.469808 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.470904 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-hxj6z" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.490039 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.491034 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.497233 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-7vnzx" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.506523 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.519868 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q498d\" (UniqueName: \"kubernetes.io/projected/cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5-kube-api-access-q498d\") pod \"placement-operator-controller-manager-78f8948974-848rm\" (UID: \"cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.519907 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.519991 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twml8\" (UniqueName: \"kubernetes.io/projected/b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad-kube-api-access-twml8\") pod \"swift-operator-controller-manager-5f8c65bbfc-ntg6w\" (UID: \"b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.520037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vnmk\" (UniqueName: \"kubernetes.io/projected/83c42810-1f8a-4616-9544-db48c6020c85-kube-api-access-2vnmk\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.520077 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfl46\" (UniqueName: \"kubernetes.io/projected/102b5fb4-119f-41ba-845e-1b691b08eb57-kube-api-access-mfl46\") pod \"ovn-operator-controller-manager-b6456fdb6-fjx5c\" (UID: \"102b5fb4-119f-41ba-845e-1b691b08eb57\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.521034 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.521085 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert podName:83c42810-1f8a-4616-9544-db48c6020c85 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:56.02106836 +0000 UTC m=+848.010126281 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" (UID: "83c42810-1f8a-4616-9544-db48c6020c85") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.525639 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.550488 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vnmk\" (UniqueName: \"kubernetes.io/projected/83c42810-1f8a-4616-9544-db48c6020c85-kube-api-access-2vnmk\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.550948 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q498d\" (UniqueName: \"kubernetes.io/projected/cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5-kube-api-access-q498d\") pod \"placement-operator-controller-manager-78f8948974-848rm\" (UID: \"cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.569073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfl46\" (UniqueName: \"kubernetes.io/projected/102b5fb4-119f-41ba-845e-1b691b08eb57-kube-api-access-mfl46\") pod \"ovn-operator-controller-manager-b6456fdb6-fjx5c\" (UID: \"102b5fb4-119f-41ba-845e-1b691b08eb57\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.628360 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.634102 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dkgw\" (UniqueName: \"kubernetes.io/projected/d602d407-4c7d-4057-a87b-b7a6fb426cfe-kube-api-access-2dkgw\") pod \"test-operator-controller-manager-5854674fcc-bxnxg\" (UID: \"d602d407-4c7d-4057-a87b-b7a6fb426cfe\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.634157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nrhr\" (UniqueName: \"kubernetes.io/projected/7540e715-d2fd-4d78-a64c-e970f3da3c9b-kube-api-access-7nrhr\") pod \"telemetry-operator-controller-manager-79bd889cc6-cprkk\" (UID: \"7540e715-d2fd-4d78-a64c-e970f3da3c9b\") " pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.634346 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twml8\" (UniqueName: \"kubernetes.io/projected/b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad-kube-api-access-twml8\") pod \"swift-operator-controller-manager-5f8c65bbfc-ntg6w\" (UID: \"b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.644210 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.647555 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.654112 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twml8\" (UniqueName: \"kubernetes.io/projected/b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad-kube-api-access-twml8\") pod \"swift-operator-controller-manager-5f8c65bbfc-ntg6w\" (UID: \"b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.667055 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-gh4sd" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.674826 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.687355 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.715509 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.733053 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.734016 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.735909 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.735997 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-cp457" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.736191 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.736394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.736473 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm8hx\" (UniqueName: \"kubernetes.io/projected/3eb6d34f-5dfb-4eb6-8d81-a7f11406531f-kube-api-access-sm8hx\") pod \"watcher-operator-controller-manager-769dc69bc-qrbr4\" (UID: \"3eb6d34f-5dfb-4eb6-8d81-a7f11406531f\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.736526 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dkgw\" (UniqueName: \"kubernetes.io/projected/d602d407-4c7d-4057-a87b-b7a6fb426cfe-kube-api-access-2dkgw\") pod \"test-operator-controller-manager-5854674fcc-bxnxg\" (UID: \"d602d407-4c7d-4057-a87b-b7a6fb426cfe\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.736546 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nrhr\" (UniqueName: \"kubernetes.io/projected/7540e715-d2fd-4d78-a64c-e970f3da3c9b-kube-api-access-7nrhr\") pod \"telemetry-operator-controller-manager-79bd889cc6-cprkk\" (UID: \"7540e715-d2fd-4d78-a64c-e970f3da3c9b\") " pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.736793 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.736833 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert podName:ceb61e3f-5534-46ad-ba39-1fe0b07530f7 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:56.736820809 +0000 UTC m=+848.725878730 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert") pod "infra-operator-controller-manager-57548d458d-dgxbb" (UID: "ceb61e3f-5534-46ad-ba39-1fe0b07530f7") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.747527 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.753000 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nrhr\" (UniqueName: \"kubernetes.io/projected/7540e715-d2fd-4d78-a64c-e970f3da3c9b-kube-api-access-7nrhr\") pod \"telemetry-operator-controller-manager-79bd889cc6-cprkk\" (UID: \"7540e715-d2fd-4d78-a64c-e970f3da3c9b\") " pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.755970 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dkgw\" (UniqueName: \"kubernetes.io/projected/d602d407-4c7d-4057-a87b-b7a6fb426cfe-kube-api-access-2dkgw\") pod \"test-operator-controller-manager-5854674fcc-bxnxg\" (UID: \"d602d407-4c7d-4057-a87b-b7a6fb426cfe\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.761866 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.762993 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.768044 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-dh8f6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.770295 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.781271 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.802914 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.829208 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.838659 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm8hx\" (UniqueName: \"kubernetes.io/projected/3eb6d34f-5dfb-4eb6-8d81-a7f11406531f-kube-api-access-sm8hx\") pod \"watcher-operator-controller-manager-769dc69bc-qrbr4\" (UID: \"3eb6d34f-5dfb-4eb6-8d81-a7f11406531f\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.839494 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr6bw\" (UniqueName: \"kubernetes.io/projected/717a3535-87a4-4d0a-b76d-a540a0cefa90-kube-api-access-cr6bw\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.839597 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.839727 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8cgm\" (UniqueName: \"kubernetes.io/projected/c0a2f26e-7dc2-444f-aa8a-75e122512c98-kube-api-access-z8cgm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-ksvz2\" (UID: \"c0a2f26e-7dc2-444f-aa8a-75e122512c98\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.839766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.842618 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m"] Dec 01 08:31:55 crc kubenswrapper[4744]: W1201 08:31:55.857516 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bc8fb90_3b43_43f8_8df4_128c639ad76a.slice/crio-019c2b18c0fb2ec1fd9e8a2e66735cb44cdf06bda26308c37a2152dc393abb94 WatchSource:0}: Error finding container 019c2b18c0fb2ec1fd9e8a2e66735cb44cdf06bda26308c37a2152dc393abb94: Status 404 returned error can't find the container with id 019c2b18c0fb2ec1fd9e8a2e66735cb44cdf06bda26308c37a2152dc393abb94 Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.860916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm8hx\" (UniqueName: \"kubernetes.io/projected/3eb6d34f-5dfb-4eb6-8d81-a7f11406531f-kube-api-access-sm8hx\") pod \"watcher-operator-controller-manager-769dc69bc-qrbr4\" (UID: \"3eb6d34f-5dfb-4eb6-8d81-a7f11406531f\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.873198 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.941340 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.941424 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8cgm\" (UniqueName: \"kubernetes.io/projected/c0a2f26e-7dc2-444f-aa8a-75e122512c98-kube-api-access-z8cgm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-ksvz2\" (UID: \"c0a2f26e-7dc2-444f-aa8a-75e122512c98\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.941448 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.941522 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr6bw\" (UniqueName: \"kubernetes.io/projected/717a3535-87a4-4d0a-b76d-a540a0cefa90-kube-api-access-cr6bw\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.941834 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.941877 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:56.441865138 +0000 UTC m=+848.430923059 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "webhook-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.942124 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: E1201 08:31:55.942151 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:56.442143745 +0000 UTC m=+848.431201666 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "metrics-server-cert" not found Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.966643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr6bw\" (UniqueName: \"kubernetes.io/projected/717a3535-87a4-4d0a-b76d-a540a0cefa90-kube-api-access-cr6bw\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.970337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8cgm\" (UniqueName: \"kubernetes.io/projected/c0a2f26e-7dc2-444f-aa8a-75e122512c98-kube-api-access-z8cgm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-ksvz2\" (UID: \"c0a2f26e-7dc2-444f-aa8a-75e122512c98\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.971852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" event={"ID":"7bc8fb90-3b43-43f8-8df4-128c639ad76a","Type":"ContainerStarted","Data":"019c2b18c0fb2ec1fd9e8a2e66735cb44cdf06bda26308c37a2152dc393abb94"} Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.983540 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp"] Dec 01 08:31:55 crc kubenswrapper[4744]: I1201 08:31:55.991389 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.019833 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9f50601_c7cd_4db6_82dd_9581a5dc8a10.slice/crio-7e7e760bed9d50264bc7c36d821f8ce75a54bf73020e9702398b00dab7ff6435 WatchSource:0}: Error finding container 7e7e760bed9d50264bc7c36d821f8ce75a54bf73020e9702398b00dab7ff6435: Status 404 returned error can't find the container with id 7e7e760bed9d50264bc7c36d821f8ce75a54bf73020e9702398b00dab7ff6435 Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.044612 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.044754 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.044823 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert podName:83c42810-1f8a-4616-9544-db48c6020c85 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:57.044807825 +0000 UTC m=+849.033865746 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" (UID: "83c42810-1f8a-4616-9544-db48c6020c85") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.090633 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.407106 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.412749 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6"] Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.417914 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5fa46a8_32d1_4eda_89ec_04bdfd35ea78.slice/crio-3af043cbbb967fb22eb9a3b65a55cda607cdcfe06350a4c19a609ce807914906 WatchSource:0}: Error finding container 3af043cbbb967fb22eb9a3b65a55cda607cdcfe06350a4c19a609ce807914906: Status 404 returned error can't find the container with id 3af043cbbb967fb22eb9a3b65a55cda607cdcfe06350a4c19a609ce807914906 Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.418698 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd"] Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.419656 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f4c0ab8_4dee_4c39_b630_061580e95592.slice/crio-cf04a29130d1b129914bed7e1e15621e1648e525e012c03e9595cec8692a1dec WatchSource:0}: Error finding container cf04a29130d1b129914bed7e1e15621e1648e525e012c03e9595cec8692a1dec: Status 404 returned error can't find the container with id cf04a29130d1b129914bed7e1e15621e1648e525e012c03e9595cec8692a1dec Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.443560 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.454892 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.454996 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.457482 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv"] Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.457563 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.457619 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:57.457604151 +0000 UTC m=+849.446662072 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "webhook-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.457897 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.457960 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:57.45793988 +0000 UTC m=+849.446997901 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "metrics-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.463370 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.522201 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-htgsw"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.528308 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.533779 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh"] Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.537092 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cf33df0_316f_4acf_b105_8dc967fae633.slice/crio-fdc6d78eddc1ef8642ef67c54d7744801f02623dcc0a147652f59db2a1b60efe WatchSource:0}: Error finding container fdc6d78eddc1ef8642ef67c54d7744801f02623dcc0a147652f59db2a1b60efe: Status 404 returned error can't find the container with id fdc6d78eddc1ef8642ef67c54d7744801f02623dcc0a147652f59db2a1b60efe Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.538958 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e8a8cd5_653f_4ef0_97c6_36b1a045f59a.slice/crio-f5cdac04b395d7f9af6477960dcbd856fdf0ea2033806840c6e1c19da0964b0d WatchSource:0}: Error finding container f5cdac04b395d7f9af6477960dcbd856fdf0ea2033806840c6e1c19da0964b0d: Status 404 returned error can't find the container with id f5cdac04b395d7f9af6477960dcbd856fdf0ea2033806840c6e1c19da0964b0d Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.538961 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh"] Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.540508 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0f26427_9c16_40b5_ad62_5a7161cba83c.slice/crio-f9ed49128c5c67b8dab30d752e464f1a4d96dd4d86d9f2a66355b10c73b4f37b WatchSource:0}: Error finding container f9ed49128c5c67b8dab30d752e464f1a4d96dd4d86d9f2a66355b10c73b4f37b: Status 404 returned error can't find the container with id f9ed49128c5c67b8dab30d752e464f1a4d96dd4d86d9f2a66355b10c73b4f37b Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.670348 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.684801 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.689515 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2"] Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.700430 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0a2f26e_7dc2_444f_aa8a_75e122512c98.slice/crio-11dc9fdb9b5caef382ab8a63e86450dcd600240f1ee5809f563d1c515d978963 WatchSource:0}: Error finding container 11dc9fdb9b5caef382ab8a63e86450dcd600240f1ee5809f563d1c515d978963: Status 404 returned error can't find the container with id 11dc9fdb9b5caef382ab8a63e86450dcd600240f1ee5809f563d1c515d978963 Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.700618 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7540e715_d2fd_4d78_a64c_e970f3da3c9b.slice/crio-e0f2eb49b99ca787aed8dd51114adee32764cb9cf356ff7f25b9da776a0000e8 WatchSource:0}: Error finding container e0f2eb49b99ca787aed8dd51114adee32764cb9cf356ff7f25b9da776a0000e8: Status 404 returned error can't find the container with id e0f2eb49b99ca787aed8dd51114adee32764cb9cf356ff7f25b9da776a0000e8 Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.703444 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-848rm"] Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.703772 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z8cgm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-ksvz2_openstack-operators(c0a2f26e-7dc2-444f-aa8a-75e122512c98): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.703899 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.12:5001/openstack-k8s-operators/telemetry-operator:4b9b5976885dec7b8bba09fe9749f3929a03aa17,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7nrhr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-79bd889cc6-cprkk_openstack-operators(7540e715-d2fd-4d78-a64c-e970f3da3c9b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.706102 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" podUID="c0a2f26e-7dc2-444f-aa8a-75e122512c98" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.707182 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7nrhr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-79bd889cc6-cprkk_openstack-operators(7540e715-d2fd-4d78-a64c-e970f3da3c9b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.708599 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" podUID="7540e715-d2fd-4d78-a64c-e970f3da3c9b" Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.721459 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96"] Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.729384 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sm8hx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-qrbr4_openstack-operators(3eb6d34f-5dfb-4eb6-8d81-a7f11406531f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.729620 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q498d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-848rm_openstack-operators(cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.729773 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r58g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-rbt96_openstack-operators(b42166df-9137-4ddb-b9b0-33515fcab298): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.731644 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r58g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-rbt96_openstack-operators(b42166df-9137-4ddb-b9b0-33515fcab298): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.733290 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" podUID="b42166df-9137-4ddb-b9b0-33515fcab298" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.736540 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sm8hx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-qrbr4_openstack-operators(3eb6d34f-5dfb-4eb6-8d81-a7f11406531f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.737825 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" podUID="3eb6d34f-5dfb-4eb6-8d81-a7f11406531f" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.736393 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q498d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-848rm_openstack-operators(cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.740447 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" podUID="cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.744374 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twml8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-ntg6w_openstack-operators(b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.747760 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w"] Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.748423 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twml8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-ntg6w_openstack-operators(b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.749753 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" podUID="b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad" Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.756448 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4"] Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.763258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.763526 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.763567 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert podName:ceb61e3f-5534-46ad-ba39-1fe0b07530f7 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:58.76355516 +0000 UTC m=+850.752613081 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert") pod "infra-operator-controller-manager-57548d458d-dgxbb" (UID: "ceb61e3f-5534-46ad-ba39-1fe0b07530f7") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.854217 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c"] Dec 01 08:31:56 crc kubenswrapper[4744]: W1201 08:31:56.856811 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod102b5fb4_119f_41ba_845e_1b691b08eb57.slice/crio-2d42b3c87fbc8d38eec03612cdb646ab002302cff4c2c196d35d1953c75680ec WatchSource:0}: Error finding container 2d42b3c87fbc8d38eec03612cdb646ab002302cff4c2c196d35d1953c75680ec: Status 404 returned error can't find the container with id 2d42b3c87fbc8d38eec03612cdb646ab002302cff4c2c196d35d1953c75680ec Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.983810 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" event={"ID":"2fbe6884-845e-4c7d-aff5-0deb483f4d19","Type":"ContainerStarted","Data":"2d115af33ece240539cfd638cd81f209f95cb6dfce97baabe71381356c3743a0"} Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.984621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" event={"ID":"214d7853-ed12-4624-82cd-6b279a827a13","Type":"ContainerStarted","Data":"7b1d28ac9125fbcaf68bc30b536f6bd53ce92ae8d0385d1533a3ac3002708974"} Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.985567 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" event={"ID":"cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5","Type":"ContainerStarted","Data":"f9b933de9fa19bea74dbb9bb11b3b883df20cdf0764086329f7c16c28255c67d"} Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.987301 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" event={"ID":"c0f26427-9c16-40b5-ad62-5a7161cba83c","Type":"ContainerStarted","Data":"f9ed49128c5c67b8dab30d752e464f1a4d96dd4d86d9f2a66355b10c73b4f37b"} Dec 01 08:31:56 crc kubenswrapper[4744]: I1201 08:31:56.998429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" event={"ID":"f5fa46a8-32d1-4eda-89ec-04bdfd35ea78","Type":"ContainerStarted","Data":"3af043cbbb967fb22eb9a3b65a55cda607cdcfe06350a4c19a609ce807914906"} Dec 01 08:31:56 crc kubenswrapper[4744]: E1201 08:31:56.998546 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" podUID="cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5" Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.000554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" event={"ID":"c0a2f26e-7dc2-444f-aa8a-75e122512c98","Type":"ContainerStarted","Data":"11dc9fdb9b5caef382ab8a63e86450dcd600240f1ee5809f563d1c515d978963"} Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.002338 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" podUID="c0a2f26e-7dc2-444f-aa8a-75e122512c98" Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.011192 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" event={"ID":"102b5fb4-119f-41ba-845e-1b691b08eb57","Type":"ContainerStarted","Data":"2d42b3c87fbc8d38eec03612cdb646ab002302cff4c2c196d35d1953c75680ec"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.013414 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" event={"ID":"d602d407-4c7d-4057-a87b-b7a6fb426cfe","Type":"ContainerStarted","Data":"eb5a8f083b22957f66697c95a5356ecf5ac3d80ff854c89e2ebb1115a02ddcc6"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.014334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" event={"ID":"7540e715-d2fd-4d78-a64c-e970f3da3c9b","Type":"ContainerStarted","Data":"e0f2eb49b99ca787aed8dd51114adee32764cb9cf356ff7f25b9da776a0000e8"} Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.016600 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/openstack-k8s-operators/telemetry-operator:4b9b5976885dec7b8bba09fe9749f3929a03aa17\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" podUID="7540e715-d2fd-4d78-a64c-e970f3da3c9b" Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.018783 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" event={"ID":"5f4c0ab8-4dee-4c39-b630-061580e95592","Type":"ContainerStarted","Data":"cf04a29130d1b129914bed7e1e15621e1648e525e012c03e9595cec8692a1dec"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.024444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" event={"ID":"4cf33df0-316f-4acf-b105-8dc967fae633","Type":"ContainerStarted","Data":"fdc6d78eddc1ef8642ef67c54d7744801f02623dcc0a147652f59db2a1b60efe"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.025470 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" event={"ID":"f68be396-2f8f-4c90-99d3-bbe12b7d6c8e","Type":"ContainerStarted","Data":"ee47bd031c3e1ea87ca895b60286c76cadf1e66cce52d87f0872d5625896c77f"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.030136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" event={"ID":"b42166df-9137-4ddb-b9b0-33515fcab298","Type":"ContainerStarted","Data":"67f72af73b9757147a58c05fb93cc4976ad3f877151fd90c6ecf4950d56ce509"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.033852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" event={"ID":"badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a","Type":"ContainerStarted","Data":"3a4ebf9924f8683ad9f22f22c0884845438928e957db96a58d780491aa94b9c7"} Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.033845 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" podUID="b42166df-9137-4ddb-b9b0-33515fcab298" Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.035684 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" event={"ID":"3eb6d34f-5dfb-4eb6-8d81-a7f11406531f","Type":"ContainerStarted","Data":"579b9cdd0d07554535df8822700a3888ba954338c3c313d180ae5c35a6c55c6a"} Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.043874 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" podUID="3eb6d34f-5dfb-4eb6-8d81-a7f11406531f" Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.046162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" event={"ID":"d9f50601-c7cd-4db6-82dd-9581a5dc8a10","Type":"ContainerStarted","Data":"7e7e760bed9d50264bc7c36d821f8ce75a54bf73020e9702398b00dab7ff6435"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.047219 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" event={"ID":"b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad","Type":"ContainerStarted","Data":"9bfe139a71ed409755f510172c5f3948bfd08ca43b79c08ac51eb66b66c6be24"} Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.049829 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" podUID="b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad" Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.050358 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" event={"ID":"5e8a8cd5-653f-4ef0-97c6-36b1a045f59a","Type":"ContainerStarted","Data":"f5cdac04b395d7f9af6477960dcbd856fdf0ea2033806840c6e1c19da0964b0d"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.051363 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" event={"ID":"8b89e5df-c38d-4465-aa74-b7c67e268a0d","Type":"ContainerStarted","Data":"a4cae4f5b6fafad49d5bd50cb3b43a22ed4e98af55c2091d903f71dceff71012"} Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.066339 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.066523 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.066745 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert podName:83c42810-1f8a-4616-9544-db48c6020c85 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:59.066724961 +0000 UTC m=+851.055782952 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" (UID: "83c42810-1f8a-4616-9544-db48c6020c85") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.473013 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:57 crc kubenswrapper[4744]: I1201 08:31:57.473110 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.473171 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.473227 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:59.47321094 +0000 UTC m=+851.462268861 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "webhook-server-cert" not found Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.473235 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:31:57 crc kubenswrapper[4744]: E1201 08:31:57.473264 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:31:59.473256411 +0000 UTC m=+851.462314332 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "metrics-server-cert" not found Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.072366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" podUID="c0a2f26e-7dc2-444f-aa8a-75e122512c98" Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.073197 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.12:5001/openstack-k8s-operators/telemetry-operator:4b9b5976885dec7b8bba09fe9749f3929a03aa17\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" podUID="7540e715-d2fd-4d78-a64c-e970f3da3c9b" Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.073574 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" podUID="3eb6d34f-5dfb-4eb6-8d81-a7f11406531f" Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.075228 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" podUID="cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5" Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.076005 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" podUID="b42166df-9137-4ddb-b9b0-33515fcab298" Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.076529 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" podUID="b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad" Dec 01 08:31:58 crc kubenswrapper[4744]: I1201 08:31:58.789626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.789947 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:58 crc kubenswrapper[4744]: E1201 08:31:58.790017 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert podName:ceb61e3f-5534-46ad-ba39-1fe0b07530f7 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:02.789998926 +0000 UTC m=+854.779056847 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert") pod "infra-operator-controller-manager-57548d458d-dgxbb" (UID: "ceb61e3f-5534-46ad-ba39-1fe0b07530f7") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:31:59 crc kubenswrapper[4744]: I1201 08:31:59.093930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:31:59 crc kubenswrapper[4744]: E1201 08:31:59.094133 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:59 crc kubenswrapper[4744]: E1201 08:31:59.094178 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert podName:83c42810-1f8a-4616-9544-db48c6020c85 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:03.094164195 +0000 UTC m=+855.083222116 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" (UID: "83c42810-1f8a-4616-9544-db48c6020c85") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:31:59 crc kubenswrapper[4744]: I1201 08:31:59.500633 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:59 crc kubenswrapper[4744]: I1201 08:31:59.500742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:31:59 crc kubenswrapper[4744]: E1201 08:31:59.500934 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:31:59 crc kubenswrapper[4744]: E1201 08:31:59.500951 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:31:59 crc kubenswrapper[4744]: E1201 08:31:59.500999 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:03.500978452 +0000 UTC m=+855.490036373 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "metrics-server-cert" not found Dec 01 08:31:59 crc kubenswrapper[4744]: E1201 08:31:59.501104 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:03.501038623 +0000 UTC m=+855.490096544 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "webhook-server-cert" not found Dec 01 08:32:02 crc kubenswrapper[4744]: I1201 08:32:02.862736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:32:02 crc kubenswrapper[4744]: E1201 08:32:02.862918 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 08:32:02 crc kubenswrapper[4744]: E1201 08:32:02.863181 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert podName:ceb61e3f-5534-46ad-ba39-1fe0b07530f7 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:10.863163622 +0000 UTC m=+862.852221543 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert") pod "infra-operator-controller-manager-57548d458d-dgxbb" (UID: "ceb61e3f-5534-46ad-ba39-1fe0b07530f7") : secret "infra-operator-webhook-server-cert" not found Dec 01 08:32:03 crc kubenswrapper[4744]: I1201 08:32:03.166655 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:32:03 crc kubenswrapper[4744]: E1201 08:32:03.166928 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:32:03 crc kubenswrapper[4744]: E1201 08:32:03.167045 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert podName:83c42810-1f8a-4616-9544-db48c6020c85 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:11.167018582 +0000 UTC m=+863.156076533 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" (UID: "83c42810-1f8a-4616-9544-db48c6020c85") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 08:32:03 crc kubenswrapper[4744]: I1201 08:32:03.573260 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:03 crc kubenswrapper[4744]: I1201 08:32:03.573382 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:03 crc kubenswrapper[4744]: E1201 08:32:03.573560 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:32:03 crc kubenswrapper[4744]: E1201 08:32:03.573683 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:11.573655764 +0000 UTC m=+863.562713715 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "webhook-server-cert" not found Dec 01 08:32:03 crc kubenswrapper[4744]: E1201 08:32:03.573569 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 08:32:03 crc kubenswrapper[4744]: E1201 08:32:03.573756 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:11.573741917 +0000 UTC m=+863.562799878 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "metrics-server-cert" not found Dec 01 08:32:09 crc kubenswrapper[4744]: I1201 08:32:09.158148 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" event={"ID":"8b89e5df-c38d-4465-aa74-b7c67e268a0d","Type":"ContainerStarted","Data":"8f53d92cec03642c4765617300670b1644c06e733f372c5643f1b7672c153d28"} Dec 01 08:32:09 crc kubenswrapper[4744]: I1201 08:32:09.161366 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" event={"ID":"f5fa46a8-32d1-4eda-89ec-04bdfd35ea78","Type":"ContainerStarted","Data":"d043a777c3ce98180471292bbe752e3a3e384b4f35402949e770450504331519"} Dec 01 08:32:09 crc kubenswrapper[4744]: I1201 08:32:09.162703 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" event={"ID":"d9f50601-c7cd-4db6-82dd-9581a5dc8a10","Type":"ContainerStarted","Data":"26c6c31d0ec6eb08a007602763f9da96a41fe8e6fa61c643657cb1276c4852f8"} Dec 01 08:32:09 crc kubenswrapper[4744]: I1201 08:32:09.165634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" event={"ID":"f68be396-2f8f-4c90-99d3-bbe12b7d6c8e","Type":"ContainerStarted","Data":"6537b6d91f0b7d6abe56a4cfb8cee3a635554779e5dbca0115fe3e0b7a5f2a1a"} Dec 01 08:32:09 crc kubenswrapper[4744]: I1201 08:32:09.173536 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" event={"ID":"102b5fb4-119f-41ba-845e-1b691b08eb57","Type":"ContainerStarted","Data":"3d1b5a377edeef0ab97ad1662af95b9046d12d8a6b6a7ef59db2f732b3e6e90a"} Dec 01 08:32:09 crc kubenswrapper[4744]: I1201 08:32:09.177469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" event={"ID":"214d7853-ed12-4624-82cd-6b279a827a13","Type":"ContainerStarted","Data":"824142eb2d99d2f42385c0077c84e53190a76b33004cd3c3db77edfacc558a4e"} Dec 01 08:32:09 crc kubenswrapper[4744]: I1201 08:32:09.180027 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" event={"ID":"c0f26427-9c16-40b5-ad62-5a7161cba83c","Type":"ContainerStarted","Data":"f34c7900070eebf58c82ce600178a3224b270e476873eadc6904da3ef20cb770"} Dec 01 08:32:09 crc kubenswrapper[4744]: E1201 08:32:09.228055 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47nmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-7z2ml_openstack-operators(badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:32:09 crc kubenswrapper[4744]: E1201 08:32:09.228169 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jcwqd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-8zfjd_openstack-operators(2fbe6884-845e-4c7d-aff5-0deb483f4d19): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:32:09 crc kubenswrapper[4744]: E1201 08:32:09.229861 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" podUID="2fbe6884-845e-4c7d-aff5-0deb483f4d19" Dec 01 08:32:09 crc kubenswrapper[4744]: E1201 08:32:09.229932 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" podUID="badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a" Dec 01 08:32:09 crc kubenswrapper[4744]: E1201 08:32:09.236349 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2dkgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-bxnxg_openstack-operators(d602d407-4c7d-4057-a87b-b7a6fb426cfe): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 08:32:09 crc kubenswrapper[4744]: E1201 08:32:09.238244 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" podUID="d602d407-4c7d-4057-a87b-b7a6fb426cfe" Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.197187 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" event={"ID":"badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a","Type":"ContainerStarted","Data":"4ea9be67a60dc9f40b547b2de629a2f70df58077bf0315bae4196aacfc728246"} Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.197971 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" Dec 01 08:32:10 crc kubenswrapper[4744]: E1201 08:32:10.200133 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" podUID="badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a" Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.212493 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" event={"ID":"5f4c0ab8-4dee-4c39-b630-061580e95592","Type":"ContainerStarted","Data":"359fe545867f6a5d246286c98f0eeab763c77ceb98e3f1e2e657cbeb5573a9c5"} Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.214564 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" event={"ID":"4cf33df0-316f-4acf-b105-8dc967fae633","Type":"ContainerStarted","Data":"0aded3d4aba9c1e003960d513bbcb4a3086c48f3edccb2d9eeec17b65c8db8b5"} Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.234360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" event={"ID":"5e8a8cd5-653f-4ef0-97c6-36b1a045f59a","Type":"ContainerStarted","Data":"0f0a6ebb0440f20830914975ef832aed03fda13fb4485d5be5a59e8f42804fd1"} Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.237573 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" event={"ID":"7bc8fb90-3b43-43f8-8df4-128c639ad76a","Type":"ContainerStarted","Data":"fa9d8e71c37aaf93df5887f594e9ac4aeb689b504fe68d0e6ba312c9a2113cfa"} Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.239183 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" event={"ID":"d602d407-4c7d-4057-a87b-b7a6fb426cfe","Type":"ContainerStarted","Data":"699019ff993c89caf55348a8d26e481903ed23861fe514454c045ea11041d89f"} Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.239491 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.240995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" event={"ID":"2fbe6884-845e-4c7d-aff5-0deb483f4d19","Type":"ContainerStarted","Data":"1adea1c21ce06fa9aee6be0498b6c3785c971d2351defbe78302012c17f6e367"} Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.241731 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" Dec 01 08:32:10 crc kubenswrapper[4744]: E1201 08:32:10.242367 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" podUID="2fbe6884-845e-4c7d-aff5-0deb483f4d19" Dec 01 08:32:10 crc kubenswrapper[4744]: E1201 08:32:10.242375 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" podUID="d602d407-4c7d-4057-a87b-b7a6fb426cfe" Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.910428 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:32:10 crc kubenswrapper[4744]: I1201 08:32:10.919883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ceb61e3f-5534-46ad-ba39-1fe0b07530f7-cert\") pod \"infra-operator-controller-manager-57548d458d-dgxbb\" (UID: \"ceb61e3f-5534-46ad-ba39-1fe0b07530f7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:32:11 crc kubenswrapper[4744]: I1201 08:32:11.165061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:32:11 crc kubenswrapper[4744]: I1201 08:32:11.220542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:32:11 crc kubenswrapper[4744]: I1201 08:32:11.224731 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83c42810-1f8a-4616-9544-db48c6020c85-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4khmps\" (UID: \"83c42810-1f8a-4616-9544-db48c6020c85\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:32:11 crc kubenswrapper[4744]: E1201 08:32:11.247544 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" podUID="badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a" Dec 01 08:32:11 crc kubenswrapper[4744]: E1201 08:32:11.247585 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" podUID="2fbe6884-845e-4c7d-aff5-0deb483f4d19" Dec 01 08:32:11 crc kubenswrapper[4744]: E1201 08:32:11.247613 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" podUID="d602d407-4c7d-4057-a87b-b7a6fb426cfe" Dec 01 08:32:11 crc kubenswrapper[4744]: I1201 08:32:11.339781 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:32:11 crc kubenswrapper[4744]: I1201 08:32:11.626256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:11 crc kubenswrapper[4744]: I1201 08:32:11.626319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:11 crc kubenswrapper[4744]: E1201 08:32:11.626379 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 08:32:11 crc kubenswrapper[4744]: E1201 08:32:11.626503 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs podName:717a3535-87a4-4d0a-b76d-a540a0cefa90 nodeName:}" failed. No retries permitted until 2025-12-01 08:32:27.626486733 +0000 UTC m=+879.615544654 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs") pod "openstack-operator-controller-manager-76f5b5884f-78kk6" (UID: "717a3535-87a4-4d0a-b76d-a540a0cefa90") : secret "webhook-server-cert" not found Dec 01 08:32:11 crc kubenswrapper[4744]: I1201 08:32:11.639773 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-metrics-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:13 crc kubenswrapper[4744]: I1201 08:32:13.086418 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb"] Dec 01 08:32:13 crc kubenswrapper[4744]: I1201 08:32:13.115212 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps"] Dec 01 08:32:15 crc kubenswrapper[4744]: I1201 08:32:15.227493 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" Dec 01 08:32:15 crc kubenswrapper[4744]: E1201 08:32:15.229893 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" podUID="2fbe6884-845e-4c7d-aff5-0deb483f4d19" Dec 01 08:32:15 crc kubenswrapper[4744]: I1201 08:32:15.239312 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" Dec 01 08:32:15 crc kubenswrapper[4744]: E1201 08:32:15.241815 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" podUID="badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a" Dec 01 08:32:15 crc kubenswrapper[4744]: I1201 08:32:15.280233 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" event={"ID":"ceb61e3f-5534-46ad-ba39-1fe0b07530f7","Type":"ContainerStarted","Data":"68a65edefe4fd38f23ff84f9e1bf33b67c4ab2e549387438aafba5a6729b6a99"} Dec 01 08:32:15 crc kubenswrapper[4744]: I1201 08:32:15.282565 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" event={"ID":"83c42810-1f8a-4616-9544-db48c6020c85","Type":"ContainerStarted","Data":"3c4b4d605f8f3089b2ef4f0fc90d5a86229c69c4f476f1d2292742774392b87e"} Dec 01 08:32:15 crc kubenswrapper[4744]: I1201 08:32:15.874868 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" Dec 01 08:32:15 crc kubenswrapper[4744]: E1201 08:32:15.877324 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" podUID="d602d407-4c7d-4057-a87b-b7a6fb426cfe" Dec 01 08:32:27 crc kubenswrapper[4744]: I1201 08:32:27.722087 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:27 crc kubenswrapper[4744]: I1201 08:32:27.729605 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/717a3535-87a4-4d0a-b76d-a540a0cefa90-webhook-certs\") pod \"openstack-operator-controller-manager-76f5b5884f-78kk6\" (UID: \"717a3535-87a4-4d0a-b76d-a540a0cefa90\") " pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:27 crc kubenswrapper[4744]: I1201 08:32:27.870701 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.048884 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.049623 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q498d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-848rm_openstack-operators(cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.606609 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.606782 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mzzhj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-wf77m_openstack-operators(7bc8fb90-3b43-43f8-8df4-128c639ad76a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.608224 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" podUID="7bc8fb90-3b43-43f8-8df4-128c639ad76a" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.608835 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.608916 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g4n6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-pv6f6_openstack-operators(5f4c0ab8-4dee-4c39-b630-061580e95592): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.610055 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" podUID="5f4c0ab8-4dee-4c39-b630-061580e95592" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.619354 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 01 08:32:31 crc kubenswrapper[4744]: E1201 08:32:31.620034 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r58g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-rbt96_openstack-operators(b42166df-9137-4ddb-b9b0-33515fcab298): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:32:32 crc kubenswrapper[4744]: I1201 08:32:32.414726 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" Dec 01 08:32:32 crc kubenswrapper[4744]: I1201 08:32:32.414801 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" Dec 01 08:32:32 crc kubenswrapper[4744]: I1201 08:32:32.416495 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" Dec 01 08:32:32 crc kubenswrapper[4744]: I1201 08:32:32.418561 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" Dec 01 08:32:32 crc kubenswrapper[4744]: E1201 08:32:32.543043 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 01 08:32:32 crc kubenswrapper[4744]: E1201 08:32:32.543563 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twml8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-ntg6w_openstack-operators(b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:32:33 crc kubenswrapper[4744]: I1201 08:32:33.394694 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6"] Dec 01 08:32:33 crc kubenswrapper[4744]: W1201 08:32:33.428382 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod717a3535_87a4_4d0a_b76d_a540a0cefa90.slice/crio-03b8679d1c7b3601d4b2e0ce7d0cc7e06ede60c112edcf92a6d3912d112aa4b9 WatchSource:0}: Error finding container 03b8679d1c7b3601d4b2e0ce7d0cc7e06ede60c112edcf92a6d3912d112aa4b9: Status 404 returned error can't find the container with id 03b8679d1c7b3601d4b2e0ce7d0cc7e06ede60c112edcf92a6d3912d112aa4b9 Dec 01 08:32:33 crc kubenswrapper[4744]: E1201 08:32:33.794261 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" podUID="b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.442023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" event={"ID":"f5fa46a8-32d1-4eda-89ec-04bdfd35ea78","Type":"ContainerStarted","Data":"d0f4504b216e909cb6b968c7933faadd7f5038a45dc3364a204d968498baff1e"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.443225 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.450554 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.456749 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" event={"ID":"5f4c0ab8-4dee-4c39-b630-061580e95592","Type":"ContainerStarted","Data":"cfcdddcabbe0e0d6abc7120282b440583f2f6bfa1de044062f7897bb3e5e0564"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.469479 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" event={"ID":"ceb61e3f-5534-46ad-ba39-1fe0b07530f7","Type":"ContainerStarted","Data":"e0c43a4cd0dde13c99962d1840e4c34a8a933daf28993c74ba15497b10a72b51"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.490648 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" event={"ID":"b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad","Type":"ContainerStarted","Data":"ec80e5343ca2ea33495932dc0aff4ff33dad75377de44587f6cf337b5cdeb8f6"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.496635 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-kvcrb" podStartSLOduration=3.717385912 podStartE2EDuration="40.496620049s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.419625249 +0000 UTC m=+848.408683170" lastFinishedPulling="2025-12-01 08:32:33.198859386 +0000 UTC m=+885.187917307" observedRunningTime="2025-12-01 08:32:34.466853287 +0000 UTC m=+886.455911208" watchObservedRunningTime="2025-12-01 08:32:34.496620049 +0000 UTC m=+886.485677970" Dec 01 08:32:34 crc kubenswrapper[4744]: E1201 08:32:34.507561 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" podUID="b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad" Dec 01 08:32:34 crc kubenswrapper[4744]: E1201 08:32:34.524773 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" podUID="cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.528566 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" event={"ID":"7bc8fb90-3b43-43f8-8df4-128c639ad76a","Type":"ContainerStarted","Data":"b2d4caf782df37e4da20b525e89f4993ec8151108b94a6da7b415a88b2227dad"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.530964 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pv6f6" podStartSLOduration=28.331258063 podStartE2EDuration="40.530946828s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.43110706 +0000 UTC m=+848.420165001" lastFinishedPulling="2025-12-01 08:32:08.630795845 +0000 UTC m=+860.619853766" observedRunningTime="2025-12-01 08:32:34.527365288 +0000 UTC m=+886.516423209" watchObservedRunningTime="2025-12-01 08:32:34.530946828 +0000 UTC m=+886.520004749" Dec 01 08:32:34 crc kubenswrapper[4744]: E1201 08:32:34.545970 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" podUID="b42166df-9137-4ddb-b9b0-33515fcab298" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.554528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" event={"ID":"2fbe6884-845e-4c7d-aff5-0deb483f4d19","Type":"ContainerStarted","Data":"cd5c6dcfd9725ce22220245d5a0c8a6a65fad2028e2078d6d7c827bd68541701"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.579716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" event={"ID":"c0f26427-9c16-40b5-ad62-5a7161cba83c","Type":"ContainerStarted","Data":"b48d372c8f3c6dcb6e43cec15167a946cde72609eb5559269513b100c172bd73"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.580768 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.596685 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.609301 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" event={"ID":"d9f50601-c7cd-4db6-82dd-9581a5dc8a10","Type":"ContainerStarted","Data":"a18262846f2f953602771b8320e428cb0f2ddbd8e858375379e68fc092d91451"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.610296 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.625813 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" event={"ID":"c0a2f26e-7dc2-444f-aa8a-75e122512c98","Type":"ContainerStarted","Data":"c08aaa832526d09dd78fa06a124623ba16310be4cd2be1891f123eb70cf06eae"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.626255 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.630116 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-wf77m" podStartSLOduration=27.940380619 podStartE2EDuration="40.630103589s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:55.870293208 +0000 UTC m=+847.859351129" lastFinishedPulling="2025-12-01 08:32:08.560016138 +0000 UTC m=+860.549074099" observedRunningTime="2025-12-01 08:32:34.597665673 +0000 UTC m=+886.586723594" watchObservedRunningTime="2025-12-01 08:32:34.630103589 +0000 UTC m=+886.619161510" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.632537 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-8zfjd" podStartSLOduration=3.821521063 podStartE2EDuration="40.632527917s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.461067347 +0000 UTC m=+848.450125268" lastFinishedPulling="2025-12-01 08:32:33.272074201 +0000 UTC m=+885.261132122" observedRunningTime="2025-12-01 08:32:34.629541713 +0000 UTC m=+886.618599624" watchObservedRunningTime="2025-12-01 08:32:34.632527917 +0000 UTC m=+886.621585838" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.647062 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" event={"ID":"5e8a8cd5-653f-4ef0-97c6-36b1a045f59a","Type":"ContainerStarted","Data":"a0b469f2617be4d749959285d5e43b8f1b3a135eef005fc5919c169f11355632"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.647813 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.659225 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.660270 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-htgsw" podStartSLOduration=4.092648758 podStartE2EDuration="40.660257621s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.544590141 +0000 UTC m=+848.533648072" lastFinishedPulling="2025-12-01 08:32:33.112198994 +0000 UTC m=+885.101256935" observedRunningTime="2025-12-01 08:32:34.657285668 +0000 UTC m=+886.646343579" watchObservedRunningTime="2025-12-01 08:32:34.660257621 +0000 UTC m=+886.649315542" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.667998 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" event={"ID":"717a3535-87a4-4d0a-b76d-a540a0cefa90","Type":"ContainerStarted","Data":"da5ff2e9e8952d682c361ed5e347e0cc8d442a4ed46f1d2bd09bcd28e263aad1"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.668122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" event={"ID":"717a3535-87a4-4d0a-b76d-a540a0cefa90","Type":"ContainerStarted","Data":"03b8679d1c7b3601d4b2e0ce7d0cc7e06ede60c112edcf92a6d3912d112aa4b9"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.668860 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.694454 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" event={"ID":"83c42810-1f8a-4616-9544-db48c6020c85","Type":"ContainerStarted","Data":"1d5edb35524acf85dee290e2b08f250e24f440e6d0e3c4dadf1fd5892ba05948"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.695121 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.697490 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lkltp" podStartSLOduration=3.524480171 podStartE2EDuration="40.697455411s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.025895476 +0000 UTC m=+848.014953387" lastFinishedPulling="2025-12-01 08:32:33.198870706 +0000 UTC m=+885.187928627" observedRunningTime="2025-12-01 08:32:34.685373603 +0000 UTC m=+886.674431524" watchObservedRunningTime="2025-12-01 08:32:34.697455411 +0000 UTC m=+886.686513332" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.708736 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" event={"ID":"d602d407-4c7d-4057-a87b-b7a6fb426cfe","Type":"ContainerStarted","Data":"7c5844b396603d4698ea4de688941ba528ee4f83f34983baf7d99bab1a5a098c"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.712755 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" event={"ID":"214d7853-ed12-4624-82cd-6b279a827a13","Type":"ContainerStarted","Data":"37ebec93ee8403558ad8448d5b641915cbad96200eea5ca9b3490cfffae7644d"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.713717 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.722644 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" event={"ID":"8b89e5df-c38d-4465-aa74-b7c67e268a0d","Type":"ContainerStarted","Data":"8097ad9c888d31dbecd5b1f07be010a2e03c6ad0bfacd39d9c08ff5bbf99df75"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.723021 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.735741 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.738580 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.739028 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" event={"ID":"badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a","Type":"ContainerStarted","Data":"21733d14ca0214fbc2e570fbe43bcb319ccc28b6bb9317318a23ce47164113ac"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.742325 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" event={"ID":"3eb6d34f-5dfb-4eb6-8d81-a7f11406531f","Type":"ContainerStarted","Data":"16113d47fb9dd925cf551ad0055d46d7db3cc9c2596733da96c14f18189279f2"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.742367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" event={"ID":"3eb6d34f-5dfb-4eb6-8d81-a7f11406531f","Type":"ContainerStarted","Data":"9ce5758286fcd5759ccdefe585d0a60e4fff101598e786f73506209ac7a6c311"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.743058 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.745671 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" event={"ID":"4cf33df0-316f-4acf-b105-8dc967fae633","Type":"ContainerStarted","Data":"f310411f03d371a159dfe669ee584dd05aae84a5e1700889c02cdb8627a0b1ce"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.746785 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.748256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" event={"ID":"f68be396-2f8f-4c90-99d3-bbe12b7d6c8e","Type":"ContainerStarted","Data":"e29c7e310544c5583d38f2b2f073ed4441f402017dff3f5decc1b6243c5e9a19"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.748902 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.753259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" event={"ID":"102b5fb4-119f-41ba-845e-1b691b08eb57","Type":"ContainerStarted","Data":"53b4069115b4fe347ac043e79342d2eef7dce0e78a6a0877df1500bdfc2259be"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.754560 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.756206 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.757874 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.757964 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" event={"ID":"7540e715-d2fd-4d78-a64c-e970f3da3c9b","Type":"ContainerStarted","Data":"f353025a83a08cc6e446aa83350310ce60269e84b32c1bf26d32462cc66cc223"} Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.758264 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.760442 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-gg7s7" podStartSLOduration=4.064849301 podStartE2EDuration="40.7604264s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.467172138 +0000 UTC m=+848.456230059" lastFinishedPulling="2025-12-01 08:32:33.162749237 +0000 UTC m=+885.151807158" observedRunningTime="2025-12-01 08:32:34.757589561 +0000 UTC m=+886.746647642" watchObservedRunningTime="2025-12-01 08:32:34.7604264 +0000 UTC m=+886.749484321" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.762620 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ksvz2" podStartSLOduration=4.424278559 podStartE2EDuration="39.762610602s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.703660296 +0000 UTC m=+848.692718217" lastFinishedPulling="2025-12-01 08:32:32.041992339 +0000 UTC m=+884.031050260" observedRunningTime="2025-12-01 08:32:34.736186953 +0000 UTC m=+886.725244874" watchObservedRunningTime="2025-12-01 08:32:34.762610602 +0000 UTC m=+886.751668513" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.768637 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.790882 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" podStartSLOduration=21.133337707 podStartE2EDuration="39.790869451s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:32:14.578626314 +0000 UTC m=+866.567684235" lastFinishedPulling="2025-12-01 08:32:33.236158058 +0000 UTC m=+885.225215979" observedRunningTime="2025-12-01 08:32:34.78977038 +0000 UTC m=+886.778828301" watchObservedRunningTime="2025-12-01 08:32:34.790869451 +0000 UTC m=+886.779927372" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.820724 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-bxnxg" podStartSLOduration=3.220201963 podStartE2EDuration="39.820706455s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.689637954 +0000 UTC m=+848.678695875" lastFinishedPulling="2025-12-01 08:32:33.290142436 +0000 UTC m=+885.279200367" observedRunningTime="2025-12-01 08:32:34.814915633 +0000 UTC m=+886.803973554" watchObservedRunningTime="2025-12-01 08:32:34.820706455 +0000 UTC m=+886.809764376" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.842731 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-q9dxh" podStartSLOduration=4.21183424 podStartE2EDuration="40.84271395s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.540058635 +0000 UTC m=+848.529116556" lastFinishedPulling="2025-12-01 08:32:33.170938335 +0000 UTC m=+885.159996266" observedRunningTime="2025-12-01 08:32:34.841739383 +0000 UTC m=+886.830797304" watchObservedRunningTime="2025-12-01 08:32:34.84271395 +0000 UTC m=+886.831771871" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.870541 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-7fn8v" podStartSLOduration=4.237749343 podStartE2EDuration="40.870520797s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.530065385 +0000 UTC m=+848.519123306" lastFinishedPulling="2025-12-01 08:32:33.162836829 +0000 UTC m=+885.151894760" observedRunningTime="2025-12-01 08:32:34.865911278 +0000 UTC m=+886.854969199" watchObservedRunningTime="2025-12-01 08:32:34.870520797 +0000 UTC m=+886.859578718" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.905471 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" podStartSLOduration=39.905454653 podStartE2EDuration="39.905454653s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:32:34.901848622 +0000 UTC m=+886.890906543" watchObservedRunningTime="2025-12-01 08:32:34.905454653 +0000 UTC m=+886.894512574" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.924687 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" podStartSLOduration=24.0093818 podStartE2EDuration="39.92466963s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.729257351 +0000 UTC m=+848.718315272" lastFinishedPulling="2025-12-01 08:32:12.644545161 +0000 UTC m=+864.633603102" observedRunningTime="2025-12-01 08:32:34.920393071 +0000 UTC m=+886.909450992" watchObservedRunningTime="2025-12-01 08:32:34.92466963 +0000 UTC m=+886.913727561" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.943283 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-fjx5c" podStartSLOduration=3.639115159 podStartE2EDuration="39.94326905s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.860029136 +0000 UTC m=+848.849087057" lastFinishedPulling="2025-12-01 08:32:33.164183017 +0000 UTC m=+885.153240948" observedRunningTime="2025-12-01 08:32:34.939811163 +0000 UTC m=+886.928869084" watchObservedRunningTime="2025-12-01 08:32:34.94326905 +0000 UTC m=+886.932326971" Dec 01 08:32:34 crc kubenswrapper[4744]: I1201 08:32:34.966249 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-wqvcv" podStartSLOduration=4.269165382 podStartE2EDuration="40.966237032s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.466775097 +0000 UTC m=+848.455833018" lastFinishedPulling="2025-12-01 08:32:33.163846747 +0000 UTC m=+885.152904668" observedRunningTime="2025-12-01 08:32:34.964726959 +0000 UTC m=+886.953784880" watchObservedRunningTime="2025-12-01 08:32:34.966237032 +0000 UTC m=+886.955294953" Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.000007 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-7z2ml" podStartSLOduration=4.18143277 podStartE2EDuration="40.999989255s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.466755736 +0000 UTC m=+848.455813657" lastFinishedPulling="2025-12-01 08:32:33.285312221 +0000 UTC m=+885.274370142" observedRunningTime="2025-12-01 08:32:34.996685333 +0000 UTC m=+886.985743254" watchObservedRunningTime="2025-12-01 08:32:34.999989255 +0000 UTC m=+886.989047166" Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.136754 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-tlszh" podStartSLOduration=4.511423371 podStartE2EDuration="41.136737806s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.539698065 +0000 UTC m=+848.528755986" lastFinishedPulling="2025-12-01 08:32:33.16501249 +0000 UTC m=+885.154070421" observedRunningTime="2025-12-01 08:32:35.059877528 +0000 UTC m=+887.048935449" watchObservedRunningTime="2025-12-01 08:32:35.136737806 +0000 UTC m=+887.125795727" Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.138542 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" podStartSLOduration=24.084736096 podStartE2EDuration="40.138536526s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.703654656 +0000 UTC m=+848.692712577" lastFinishedPulling="2025-12-01 08:32:12.757455076 +0000 UTC m=+864.746513007" observedRunningTime="2025-12-01 08:32:35.128029803 +0000 UTC m=+887.117087724" watchObservedRunningTime="2025-12-01 08:32:35.138536526 +0000 UTC m=+887.127594447" Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.768718 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" event={"ID":"ceb61e3f-5534-46ad-ba39-1fe0b07530f7","Type":"ContainerStarted","Data":"a5a7c4e7be12ece767b241adfb1688d94a979c3ea71ecd81ddaca8bf15beaccb"} Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.768893 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.771877 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" event={"ID":"83c42810-1f8a-4616-9544-db48c6020c85","Type":"ContainerStarted","Data":"77a33072bb257193582f432be240dde3a2d9407441887f3a24708a61480ed64f"} Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.773887 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" event={"ID":"7540e715-d2fd-4d78-a64c-e970f3da3c9b","Type":"ContainerStarted","Data":"e3f2ffa7a8a0f3492fbdc4ae85c86ba22dfe3aeb20f07b00d438ba229e7cb124"} Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.775325 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" event={"ID":"cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5","Type":"ContainerStarted","Data":"20e2b937681436e3bd007cd22ed3669eda3ab102b85b03eac94434828e7802b4"} Dec 01 08:32:35 crc kubenswrapper[4744]: E1201 08:32:35.777161 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" podUID="cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5" Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.778659 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" event={"ID":"b42166df-9137-4ddb-b9b0-33515fcab298","Type":"ContainerStarted","Data":"9c74727da28e4360fd5ae2bd36cbc3c86708eec832c0930f73a1041a77a8ee82"} Dec 01 08:32:35 crc kubenswrapper[4744]: E1201 08:32:35.783023 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" podUID="b42166df-9137-4ddb-b9b0-33515fcab298" Dec 01 08:32:35 crc kubenswrapper[4744]: I1201 08:32:35.802194 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" podStartSLOduration=23.244421323 podStartE2EDuration="41.802170489s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:32:14.578243513 +0000 UTC m=+866.567301424" lastFinishedPulling="2025-12-01 08:32:33.135992649 +0000 UTC m=+885.125050590" observedRunningTime="2025-12-01 08:32:35.797807177 +0000 UTC m=+887.786865128" watchObservedRunningTime="2025-12-01 08:32:35.802170489 +0000 UTC m=+887.791228420" Dec 01 08:32:41 crc kubenswrapper[4744]: I1201 08:32:41.177291 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-dgxbb" Dec 01 08:32:41 crc kubenswrapper[4744]: I1201 08:32:41.347162 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4khmps" Dec 01 08:32:45 crc kubenswrapper[4744]: I1201 08:32:45.836298 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-79bd889cc6-cprkk" Dec 01 08:32:45 crc kubenswrapper[4744]: I1201 08:32:45.994382 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qrbr4" Dec 01 08:32:46 crc kubenswrapper[4744]: E1201 08:32:46.287927 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" podUID="b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad" Dec 01 08:32:46 crc kubenswrapper[4744]: E1201 08:32:46.287941 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" podUID="b42166df-9137-4ddb-b9b0-33515fcab298" Dec 01 08:32:47 crc kubenswrapper[4744]: E1201 08:32:47.287206 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" podUID="cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5" Dec 01 08:32:47 crc kubenswrapper[4744]: I1201 08:32:47.879171 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-76f5b5884f-78kk6" Dec 01 08:32:56 crc kubenswrapper[4744]: I1201 08:32:56.911247 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bjxkw"] Dec 01 08:32:56 crc kubenswrapper[4744]: I1201 08:32:56.913209 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:56 crc kubenswrapper[4744]: I1201 08:32:56.920548 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bjxkw"] Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.013450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-catalog-content\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.013721 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcsfw\" (UniqueName: \"kubernetes.io/projected/af9cd3ff-578f-425a-a824-9a11c66e44d8-kube-api-access-rcsfw\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.013806 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-utilities\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.115027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-catalog-content\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.115112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcsfw\" (UniqueName: \"kubernetes.io/projected/af9cd3ff-578f-425a-a824-9a11c66e44d8-kube-api-access-rcsfw\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.115150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-utilities\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.115852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-catalog-content\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.115956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-utilities\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.141654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcsfw\" (UniqueName: \"kubernetes.io/projected/af9cd3ff-578f-425a-a824-9a11c66e44d8-kube-api-access-rcsfw\") pod \"community-operators-bjxkw\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.239015 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:32:57 crc kubenswrapper[4744]: I1201 08:32:57.722215 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bjxkw"] Dec 01 08:32:57 crc kubenswrapper[4744]: W1201 08:32:57.726135 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9cd3ff_578f_425a_a824_9a11c66e44d8.slice/crio-bcb8dbeae3a50eb11eff67e7889be9f09608e426ac5fe15ba6bdfe273eb59505 WatchSource:0}: Error finding container bcb8dbeae3a50eb11eff67e7889be9f09608e426ac5fe15ba6bdfe273eb59505: Status 404 returned error can't find the container with id bcb8dbeae3a50eb11eff67e7889be9f09608e426ac5fe15ba6bdfe273eb59505 Dec 01 08:32:58 crc kubenswrapper[4744]: I1201 08:32:58.010329 4744 generic.go:334] "Generic (PLEG): container finished" podID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerID="98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3" exitCode=0 Dec 01 08:32:58 crc kubenswrapper[4744]: I1201 08:32:58.010433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjxkw" event={"ID":"af9cd3ff-578f-425a-a824-9a11c66e44d8","Type":"ContainerDied","Data":"98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3"} Dec 01 08:32:58 crc kubenswrapper[4744]: I1201 08:32:58.010747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjxkw" event={"ID":"af9cd3ff-578f-425a-a824-9a11c66e44d8","Type":"ContainerStarted","Data":"bcb8dbeae3a50eb11eff67e7889be9f09608e426ac5fe15ba6bdfe273eb59505"} Dec 01 08:32:59 crc kubenswrapper[4744]: I1201 08:32:59.020723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjxkw" event={"ID":"af9cd3ff-578f-425a-a824-9a11c66e44d8","Type":"ContainerStarted","Data":"7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb"} Dec 01 08:33:00 crc kubenswrapper[4744]: I1201 08:33:00.033046 4744 generic.go:334] "Generic (PLEG): container finished" podID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerID="7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb" exitCode=0 Dec 01 08:33:00 crc kubenswrapper[4744]: I1201 08:33:00.033127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjxkw" event={"ID":"af9cd3ff-578f-425a-a824-9a11c66e44d8","Type":"ContainerDied","Data":"7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb"} Dec 01 08:33:01 crc kubenswrapper[4744]: I1201 08:33:01.041247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjxkw" event={"ID":"af9cd3ff-578f-425a-a824-9a11c66e44d8","Type":"ContainerStarted","Data":"d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd"} Dec 01 08:33:01 crc kubenswrapper[4744]: I1201 08:33:01.070299 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bjxkw" podStartSLOduration=2.317599279 podStartE2EDuration="5.070282264s" podCreationTimestamp="2025-12-01 08:32:56 +0000 UTC" firstStartedPulling="2025-12-01 08:32:58.013191137 +0000 UTC m=+910.002249068" lastFinishedPulling="2025-12-01 08:33:00.765874132 +0000 UTC m=+912.754932053" observedRunningTime="2025-12-01 08:33:01.068660918 +0000 UTC m=+913.057718839" watchObservedRunningTime="2025-12-01 08:33:01.070282264 +0000 UTC m=+913.059340185" Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.050272 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" event={"ID":"b42166df-9137-4ddb-b9b0-33515fcab298","Type":"ContainerStarted","Data":"679ba9f4179d7862c8e9c6a505440fddf5e27a74d592025dc4e6098d6d5fbc3a"} Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.051016 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.071816 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" podStartSLOduration=3.774315194 podStartE2EDuration="1m8.071795466s" podCreationTimestamp="2025-12-01 08:31:54 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.729701734 +0000 UTC m=+848.718759655" lastFinishedPulling="2025-12-01 08:33:01.027182006 +0000 UTC m=+913.016239927" observedRunningTime="2025-12-01 08:33:02.069879391 +0000 UTC m=+914.058937322" watchObservedRunningTime="2025-12-01 08:33:02.071795466 +0000 UTC m=+914.060853397" Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.879143 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qcgv9"] Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.880605 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.898472 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-utilities\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.898811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zlnp\" (UniqueName: \"kubernetes.io/projected/4d11be72-00bb-4df8-91b4-8e8241ad618a-kube-api-access-2zlnp\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.899012 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-catalog-content\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:02 crc kubenswrapper[4744]: I1201 08:33:02.901594 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcgv9"] Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.000175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-catalog-content\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.000512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-utilities\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.001159 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-catalog-content\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.001337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-utilities\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.001648 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zlnp\" (UniqueName: \"kubernetes.io/projected/4d11be72-00bb-4df8-91b4-8e8241ad618a-kube-api-access-2zlnp\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.028010 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zlnp\" (UniqueName: \"kubernetes.io/projected/4d11be72-00bb-4df8-91b4-8e8241ad618a-kube-api-access-2zlnp\") pod \"redhat-operators-qcgv9\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.061653 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" event={"ID":"b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad","Type":"ContainerStarted","Data":"b70a0c71393d3fe316838e76da90d9e74c1adea9110be4b0fd946555cb8b43e0"} Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.062810 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.063419 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" event={"ID":"cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5","Type":"ContainerStarted","Data":"767fb74252228c0b9384363dc8e72b05dc97ba259cc66c3d10af32beec93864f"} Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.115889 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" podStartSLOduration=2.989614836 podStartE2EDuration="1m8.115871039s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.729486558 +0000 UTC m=+848.718544479" lastFinishedPulling="2025-12-01 08:33:01.855742761 +0000 UTC m=+913.844800682" observedRunningTime="2025-12-01 08:33:03.114358066 +0000 UTC m=+915.103415987" watchObservedRunningTime="2025-12-01 08:33:03.115871039 +0000 UTC m=+915.104928960" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.118728 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" podStartSLOduration=2.893057341 podStartE2EDuration="1m8.118717431s" podCreationTimestamp="2025-12-01 08:31:55 +0000 UTC" firstStartedPulling="2025-12-01 08:31:56.744145487 +0000 UTC m=+848.733203408" lastFinishedPulling="2025-12-01 08:33:01.969805557 +0000 UTC m=+913.958863498" observedRunningTime="2025-12-01 08:33:03.093566589 +0000 UTC m=+915.082624510" watchObservedRunningTime="2025-12-01 08:33:03.118717431 +0000 UTC m=+915.107775342" Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.200086 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:03 crc kubenswrapper[4744]: W1201 08:33:03.669092 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d11be72_00bb_4df8_91b4_8e8241ad618a.slice/crio-8b8661aab4d1f072caff12c8f7b41c792f2387ab0a5fe93641f4d0cc3af60910 WatchSource:0}: Error finding container 8b8661aab4d1f072caff12c8f7b41c792f2387ab0a5fe93641f4d0cc3af60910: Status 404 returned error can't find the container with id 8b8661aab4d1f072caff12c8f7b41c792f2387ab0a5fe93641f4d0cc3af60910 Dec 01 08:33:03 crc kubenswrapper[4744]: I1201 08:33:03.669164 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcgv9"] Dec 01 08:33:04 crc kubenswrapper[4744]: I1201 08:33:04.069875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgv9" event={"ID":"4d11be72-00bb-4df8-91b4-8e8241ad618a","Type":"ContainerStarted","Data":"8b8661aab4d1f072caff12c8f7b41c792f2387ab0a5fe93641f4d0cc3af60910"} Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.272873 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q2bt9"] Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.277361 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.285287 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2bt9"] Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.435581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-catalog-content\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.435696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zrc7\" (UniqueName: \"kubernetes.io/projected/77f0f866-0e29-42b0-8637-a7eab12cd951-kube-api-access-4zrc7\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.435750 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-utilities\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.536999 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-catalog-content\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.537060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zrc7\" (UniqueName: \"kubernetes.io/projected/77f0f866-0e29-42b0-8637-a7eab12cd951-kube-api-access-4zrc7\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.537095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-utilities\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.537443 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-catalog-content\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.537490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-utilities\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.571520 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zrc7\" (UniqueName: \"kubernetes.io/projected/77f0f866-0e29-42b0-8637-a7eab12cd951-kube-api-access-4zrc7\") pod \"certified-operators-q2bt9\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.609814 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:05 crc kubenswrapper[4744]: I1201 08:33:05.782633 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" Dec 01 08:33:06 crc kubenswrapper[4744]: I1201 08:33:06.034681 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2bt9"] Dec 01 08:33:06 crc kubenswrapper[4744]: I1201 08:33:06.089842 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerID="80655c12bbba246396323ca0e90f22e1e9aa3576572d5515abba3d99cdc87d25" exitCode=0 Dec 01 08:33:06 crc kubenswrapper[4744]: I1201 08:33:06.089929 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgv9" event={"ID":"4d11be72-00bb-4df8-91b4-8e8241ad618a","Type":"ContainerDied","Data":"80655c12bbba246396323ca0e90f22e1e9aa3576572d5515abba3d99cdc87d25"} Dec 01 08:33:06 crc kubenswrapper[4744]: I1201 08:33:06.091092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2bt9" event={"ID":"77f0f866-0e29-42b0-8637-a7eab12cd951","Type":"ContainerStarted","Data":"fb61bf84631a436062120477af38c18a8784307d5415636eb2b267889f2a0740"} Dec 01 08:33:07 crc kubenswrapper[4744]: I1201 08:33:07.101352 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgv9" event={"ID":"4d11be72-00bb-4df8-91b4-8e8241ad618a","Type":"ContainerStarted","Data":"abdc9fffc03f19b101be95fa2347579c069f0000a4be4f37a3d50d6568bb6611"} Dec 01 08:33:07 crc kubenswrapper[4744]: I1201 08:33:07.104662 4744 generic.go:334] "Generic (PLEG): container finished" podID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerID="a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01" exitCode=0 Dec 01 08:33:07 crc kubenswrapper[4744]: I1201 08:33:07.104715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2bt9" event={"ID":"77f0f866-0e29-42b0-8637-a7eab12cd951","Type":"ContainerDied","Data":"a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01"} Dec 01 08:33:07 crc kubenswrapper[4744]: I1201 08:33:07.240240 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:33:07 crc kubenswrapper[4744]: I1201 08:33:07.240326 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:33:07 crc kubenswrapper[4744]: I1201 08:33:07.315973 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:33:08 crc kubenswrapper[4744]: I1201 08:33:08.115126 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerID="abdc9fffc03f19b101be95fa2347579c069f0000a4be4f37a3d50d6568bb6611" exitCode=0 Dec 01 08:33:08 crc kubenswrapper[4744]: I1201 08:33:08.115235 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgv9" event={"ID":"4d11be72-00bb-4df8-91b4-8e8241ad618a","Type":"ContainerDied","Data":"abdc9fffc03f19b101be95fa2347579c069f0000a4be4f37a3d50d6568bb6611"} Dec 01 08:33:08 crc kubenswrapper[4744]: I1201 08:33:08.175728 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:33:09 crc kubenswrapper[4744]: I1201 08:33:09.127716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgv9" event={"ID":"4d11be72-00bb-4df8-91b4-8e8241ad618a","Type":"ContainerStarted","Data":"04173ff773c21855372597767b524f31c269d1e4da36192d9582dc60baeb24bf"} Dec 01 08:33:09 crc kubenswrapper[4744]: I1201 08:33:09.131040 4744 generic.go:334] "Generic (PLEG): container finished" podID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerID="99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f" exitCode=0 Dec 01 08:33:09 crc kubenswrapper[4744]: I1201 08:33:09.131973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2bt9" event={"ID":"77f0f866-0e29-42b0-8637-a7eab12cd951","Type":"ContainerDied","Data":"99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f"} Dec 01 08:33:09 crc kubenswrapper[4744]: I1201 08:33:09.164565 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qcgv9" podStartSLOduration=4.5808739769999995 podStartE2EDuration="7.164543932s" podCreationTimestamp="2025-12-01 08:33:02 +0000 UTC" firstStartedPulling="2025-12-01 08:33:06.091164063 +0000 UTC m=+918.080221994" lastFinishedPulling="2025-12-01 08:33:08.674834028 +0000 UTC m=+920.663891949" observedRunningTime="2025-12-01 08:33:09.160933768 +0000 UTC m=+921.149991719" watchObservedRunningTime="2025-12-01 08:33:09.164543932 +0000 UTC m=+921.153601853" Dec 01 08:33:09 crc kubenswrapper[4744]: I1201 08:33:09.867946 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bjxkw"] Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.139495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2bt9" event={"ID":"77f0f866-0e29-42b0-8637-a7eab12cd951","Type":"ContainerStarted","Data":"86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9"} Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.140026 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bjxkw" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="registry-server" containerID="cri-o://d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd" gracePeriod=2 Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.170869 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q2bt9" podStartSLOduration=2.601716633 podStartE2EDuration="5.17084846s" podCreationTimestamp="2025-12-01 08:33:05 +0000 UTC" firstStartedPulling="2025-12-01 08:33:07.106868502 +0000 UTC m=+919.095926443" lastFinishedPulling="2025-12-01 08:33:09.676000349 +0000 UTC m=+921.665058270" observedRunningTime="2025-12-01 08:33:10.163929582 +0000 UTC m=+922.152987513" watchObservedRunningTime="2025-12-01 08:33:10.17084846 +0000 UTC m=+922.159906392" Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.610916 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.615579 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcsfw\" (UniqueName: \"kubernetes.io/projected/af9cd3ff-578f-425a-a824-9a11c66e44d8-kube-api-access-rcsfw\") pod \"af9cd3ff-578f-425a-a824-9a11c66e44d8\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.615654 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-utilities\") pod \"af9cd3ff-578f-425a-a824-9a11c66e44d8\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.615697 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-catalog-content\") pod \"af9cd3ff-578f-425a-a824-9a11c66e44d8\" (UID: \"af9cd3ff-578f-425a-a824-9a11c66e44d8\") " Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.616851 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-utilities" (OuterVolumeSpecName: "utilities") pod "af9cd3ff-578f-425a-a824-9a11c66e44d8" (UID: "af9cd3ff-578f-425a-a824-9a11c66e44d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.623939 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9cd3ff-578f-425a-a824-9a11c66e44d8-kube-api-access-rcsfw" (OuterVolumeSpecName: "kube-api-access-rcsfw") pod "af9cd3ff-578f-425a-a824-9a11c66e44d8" (UID: "af9cd3ff-578f-425a-a824-9a11c66e44d8"). InnerVolumeSpecName "kube-api-access-rcsfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.691533 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af9cd3ff-578f-425a-a824-9a11c66e44d8" (UID: "af9cd3ff-578f-425a-a824-9a11c66e44d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.717026 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcsfw\" (UniqueName: \"kubernetes.io/projected/af9cd3ff-578f-425a-a824-9a11c66e44d8-kube-api-access-rcsfw\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.717061 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:10 crc kubenswrapper[4744]: I1201 08:33:10.717071 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9cd3ff-578f-425a-a824-9a11c66e44d8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.149009 4744 generic.go:334] "Generic (PLEG): container finished" podID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerID="d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd" exitCode=0 Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.149104 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjxkw" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.149079 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjxkw" event={"ID":"af9cd3ff-578f-425a-a824-9a11c66e44d8","Type":"ContainerDied","Data":"d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd"} Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.149184 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjxkw" event={"ID":"af9cd3ff-578f-425a-a824-9a11c66e44d8","Type":"ContainerDied","Data":"bcb8dbeae3a50eb11eff67e7889be9f09608e426ac5fe15ba6bdfe273eb59505"} Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.149220 4744 scope.go:117] "RemoveContainer" containerID="d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.170641 4744 scope.go:117] "RemoveContainer" containerID="7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.195169 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bjxkw"] Dec 01 08:33:11 crc kubenswrapper[4744]: E1201 08:33:11.196040 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9cd3ff_578f_425a_a824_9a11c66e44d8.slice\": RecentStats: unable to find data in memory cache]" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.199927 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bjxkw"] Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.211105 4744 scope.go:117] "RemoveContainer" containerID="98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.233880 4744 scope.go:117] "RemoveContainer" containerID="d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd" Dec 01 08:33:11 crc kubenswrapper[4744]: E1201 08:33:11.234373 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd\": container with ID starting with d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd not found: ID does not exist" containerID="d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.234531 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd"} err="failed to get container status \"d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd\": rpc error: code = NotFound desc = could not find container \"d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd\": container with ID starting with d19a251d37249981912000fd5e5b93e8841f7ecc3b3a0412c4baa53a864f53fd not found: ID does not exist" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.234565 4744 scope.go:117] "RemoveContainer" containerID="7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb" Dec 01 08:33:11 crc kubenswrapper[4744]: E1201 08:33:11.234914 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb\": container with ID starting with 7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb not found: ID does not exist" containerID="7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.234990 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb"} err="failed to get container status \"7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb\": rpc error: code = NotFound desc = could not find container \"7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb\": container with ID starting with 7f6942405dc017dc82aecb66e25e90aa8ce1d562a917e471f9b7655f6c3ffddb not found: ID does not exist" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.235015 4744 scope.go:117] "RemoveContainer" containerID="98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3" Dec 01 08:33:11 crc kubenswrapper[4744]: E1201 08:33:11.237009 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3\": container with ID starting with 98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3 not found: ID does not exist" containerID="98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3" Dec 01 08:33:11 crc kubenswrapper[4744]: I1201 08:33:11.237059 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3"} err="failed to get container status \"98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3\": rpc error: code = NotFound desc = could not find container \"98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3\": container with ID starting with 98ca3dbb0f7af77f790f6a205c47685bdbb64fbc269f31fc68c759b68391cdb3 not found: ID does not exist" Dec 01 08:33:12 crc kubenswrapper[4744]: I1201 08:33:12.295159 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" path="/var/lib/kubelet/pods/af9cd3ff-578f-425a-a824-9a11c66e44d8/volumes" Dec 01 08:33:13 crc kubenswrapper[4744]: I1201 08:33:13.200494 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:13 crc kubenswrapper[4744]: I1201 08:33:13.200544 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.249710 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qcgv9" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="registry-server" probeResult="failure" output=< Dec 01 08:33:14 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 08:33:14 crc kubenswrapper[4744]: > Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.276505 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-52vz5"] Dec 01 08:33:14 crc kubenswrapper[4744]: E1201 08:33:14.276789 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="extract-utilities" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.276807 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="extract-utilities" Dec 01 08:33:14 crc kubenswrapper[4744]: E1201 08:33:14.276837 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="registry-server" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.276844 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="registry-server" Dec 01 08:33:14 crc kubenswrapper[4744]: E1201 08:33:14.276885 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="extract-content" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.276892 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="extract-content" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.277204 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9cd3ff-578f-425a-a824-9a11c66e44d8" containerName="registry-server" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.278223 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.325103 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52vz5"] Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.477126 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-catalog-content\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.477281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gn6t\" (UniqueName: \"kubernetes.io/projected/f41f9cdf-d105-427a-a4e7-fb1f2320111d-kube-api-access-5gn6t\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.477314 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-utilities\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.578754 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gn6t\" (UniqueName: \"kubernetes.io/projected/f41f9cdf-d105-427a-a4e7-fb1f2320111d-kube-api-access-5gn6t\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.579121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-utilities\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.579150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-catalog-content\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.579585 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-utilities\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.579729 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-catalog-content\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.596289 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gn6t\" (UniqueName: \"kubernetes.io/projected/f41f9cdf-d105-427a-a4e7-fb1f2320111d-kube-api-access-5gn6t\") pod \"redhat-marketplace-52vz5\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:14 crc kubenswrapper[4744]: I1201 08:33:14.616179 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.055540 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52vz5"] Dec 01 08:33:15 crc kubenswrapper[4744]: W1201 08:33:15.059711 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf41f9cdf_d105_427a_a4e7_fb1f2320111d.slice/crio-a9dd8cc27fcca5641824897e0ab3c3b3786c73a0d3b064de3eb5d1ab3a262a29 WatchSource:0}: Error finding container a9dd8cc27fcca5641824897e0ab3c3b3786c73a0d3b064de3eb5d1ab3a262a29: Status 404 returned error can't find the container with id a9dd8cc27fcca5641824897e0ab3c3b3786c73a0d3b064de3eb5d1ab3a262a29 Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.181259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52vz5" event={"ID":"f41f9cdf-d105-427a-a4e7-fb1f2320111d","Type":"ContainerStarted","Data":"a9dd8cc27fcca5641824897e0ab3c3b3786c73a0d3b064de3eb5d1ab3a262a29"} Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.530315 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rbt96" Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.610101 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.610471 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.681023 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.785480 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-848rm" Dec 01 08:33:15 crc kubenswrapper[4744]: I1201 08:33:15.807589 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-ntg6w" Dec 01 08:33:16 crc kubenswrapper[4744]: I1201 08:33:16.191180 4744 generic.go:334] "Generic (PLEG): container finished" podID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerID="87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d" exitCode=0 Dec 01 08:33:16 crc kubenswrapper[4744]: I1201 08:33:16.191301 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52vz5" event={"ID":"f41f9cdf-d105-427a-a4e7-fb1f2320111d","Type":"ContainerDied","Data":"87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d"} Dec 01 08:33:16 crc kubenswrapper[4744]: I1201 08:33:16.194845 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:33:16 crc kubenswrapper[4744]: I1201 08:33:16.261660 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:17 crc kubenswrapper[4744]: I1201 08:33:17.201153 4744 generic.go:334] "Generic (PLEG): container finished" podID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerID="1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2" exitCode=0 Dec 01 08:33:17 crc kubenswrapper[4744]: I1201 08:33:17.201252 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52vz5" event={"ID":"f41f9cdf-d105-427a-a4e7-fb1f2320111d","Type":"ContainerDied","Data":"1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2"} Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.065219 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2bt9"] Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.208373 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q2bt9" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="registry-server" containerID="cri-o://86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9" gracePeriod=2 Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.833940 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.950031 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-utilities\") pod \"77f0f866-0e29-42b0-8637-a7eab12cd951\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.950086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-catalog-content\") pod \"77f0f866-0e29-42b0-8637-a7eab12cd951\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.950119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zrc7\" (UniqueName: \"kubernetes.io/projected/77f0f866-0e29-42b0-8637-a7eab12cd951-kube-api-access-4zrc7\") pod \"77f0f866-0e29-42b0-8637-a7eab12cd951\" (UID: \"77f0f866-0e29-42b0-8637-a7eab12cd951\") " Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.951360 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-utilities" (OuterVolumeSpecName: "utilities") pod "77f0f866-0e29-42b0-8637-a7eab12cd951" (UID: "77f0f866-0e29-42b0-8637-a7eab12cd951"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:18 crc kubenswrapper[4744]: I1201 08:33:18.958594 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f0f866-0e29-42b0-8637-a7eab12cd951-kube-api-access-4zrc7" (OuterVolumeSpecName: "kube-api-access-4zrc7") pod "77f0f866-0e29-42b0-8637-a7eab12cd951" (UID: "77f0f866-0e29-42b0-8637-a7eab12cd951"). InnerVolumeSpecName "kube-api-access-4zrc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.005600 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77f0f866-0e29-42b0-8637-a7eab12cd951" (UID: "77f0f866-0e29-42b0-8637-a7eab12cd951"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.051799 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.051859 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zrc7\" (UniqueName: \"kubernetes.io/projected/77f0f866-0e29-42b0-8637-a7eab12cd951-kube-api-access-4zrc7\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.051883 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f0f866-0e29-42b0-8637-a7eab12cd951-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.217667 4744 generic.go:334] "Generic (PLEG): container finished" podID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerID="86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9" exitCode=0 Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.217742 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2bt9" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.217757 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2bt9" event={"ID":"77f0f866-0e29-42b0-8637-a7eab12cd951","Type":"ContainerDied","Data":"86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9"} Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.217803 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2bt9" event={"ID":"77f0f866-0e29-42b0-8637-a7eab12cd951","Type":"ContainerDied","Data":"fb61bf84631a436062120477af38c18a8784307d5415636eb2b267889f2a0740"} Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.217831 4744 scope.go:117] "RemoveContainer" containerID="86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.222277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52vz5" event={"ID":"f41f9cdf-d105-427a-a4e7-fb1f2320111d","Type":"ContainerStarted","Data":"eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5"} Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.237052 4744 scope.go:117] "RemoveContainer" containerID="99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.255662 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-52vz5" podStartSLOduration=2.904644679 podStartE2EDuration="5.255639024s" podCreationTimestamp="2025-12-01 08:33:14 +0000 UTC" firstStartedPulling="2025-12-01 08:33:16.194439203 +0000 UTC m=+928.183497154" lastFinishedPulling="2025-12-01 08:33:18.545433568 +0000 UTC m=+930.534491499" observedRunningTime="2025-12-01 08:33:19.254250274 +0000 UTC m=+931.243308205" watchObservedRunningTime="2025-12-01 08:33:19.255639024 +0000 UTC m=+931.244696965" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.257514 4744 scope.go:117] "RemoveContainer" containerID="a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.274272 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2bt9"] Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.285032 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q2bt9"] Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.287667 4744 scope.go:117] "RemoveContainer" containerID="86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9" Dec 01 08:33:19 crc kubenswrapper[4744]: E1201 08:33:19.288344 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9\": container with ID starting with 86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9 not found: ID does not exist" containerID="86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.288515 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9"} err="failed to get container status \"86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9\": rpc error: code = NotFound desc = could not find container \"86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9\": container with ID starting with 86b966040e2bd63742b6785a0b91cbf787bb9565a4503ff7fee612f6e9d2cbc9 not found: ID does not exist" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.288555 4744 scope.go:117] "RemoveContainer" containerID="99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f" Dec 01 08:33:19 crc kubenswrapper[4744]: E1201 08:33:19.290203 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f\": container with ID starting with 99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f not found: ID does not exist" containerID="99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.290252 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f"} err="failed to get container status \"99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f\": rpc error: code = NotFound desc = could not find container \"99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f\": container with ID starting with 99fa3117db1ecb51611440ccd43a29b16a3993be0bc5d708b19f8ca79b41ad8f not found: ID does not exist" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.290281 4744 scope.go:117] "RemoveContainer" containerID="a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01" Dec 01 08:33:19 crc kubenswrapper[4744]: E1201 08:33:19.290852 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01\": container with ID starting with a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01 not found: ID does not exist" containerID="a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01" Dec 01 08:33:19 crc kubenswrapper[4744]: I1201 08:33:19.291005 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01"} err="failed to get container status \"a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01\": rpc error: code = NotFound desc = could not find container \"a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01\": container with ID starting with a29ae60bb0d1a242ba5667af6d1cf8e4d59c9a48856013aaf4f3a07e8b0baa01 not found: ID does not exist" Dec 01 08:33:20 crc kubenswrapper[4744]: I1201 08:33:20.296801 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" path="/var/lib/kubelet/pods/77f0f866-0e29-42b0-8637-a7eab12cd951/volumes" Dec 01 08:33:23 crc kubenswrapper[4744]: I1201 08:33:23.285383 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:23 crc kubenswrapper[4744]: I1201 08:33:23.386244 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:24 crc kubenswrapper[4744]: I1201 08:33:24.272968 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcgv9"] Dec 01 08:33:24 crc kubenswrapper[4744]: I1201 08:33:24.617170 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:24 crc kubenswrapper[4744]: I1201 08:33:24.617547 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:24 crc kubenswrapper[4744]: I1201 08:33:24.693956 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:25 crc kubenswrapper[4744]: I1201 08:33:25.283857 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qcgv9" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="registry-server" containerID="cri-o://04173ff773c21855372597767b524f31c269d1e4da36192d9582dc60baeb24bf" gracePeriod=2 Dec 01 08:33:25 crc kubenswrapper[4744]: I1201 08:33:25.364169 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.295007 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerID="04173ff773c21855372597767b524f31c269d1e4da36192d9582dc60baeb24bf" exitCode=0 Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.295052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgv9" event={"ID":"4d11be72-00bb-4df8-91b4-8e8241ad618a","Type":"ContainerDied","Data":"04173ff773c21855372597767b524f31c269d1e4da36192d9582dc60baeb24bf"} Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.362486 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.463746 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zlnp\" (UniqueName: \"kubernetes.io/projected/4d11be72-00bb-4df8-91b4-8e8241ad618a-kube-api-access-2zlnp\") pod \"4d11be72-00bb-4df8-91b4-8e8241ad618a\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.464048 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-catalog-content\") pod \"4d11be72-00bb-4df8-91b4-8e8241ad618a\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.464196 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-utilities\") pod \"4d11be72-00bb-4df8-91b4-8e8241ad618a\" (UID: \"4d11be72-00bb-4df8-91b4-8e8241ad618a\") " Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.464908 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-utilities" (OuterVolumeSpecName: "utilities") pod "4d11be72-00bb-4df8-91b4-8e8241ad618a" (UID: "4d11be72-00bb-4df8-91b4-8e8241ad618a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.469596 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d11be72-00bb-4df8-91b4-8e8241ad618a-kube-api-access-2zlnp" (OuterVolumeSpecName: "kube-api-access-2zlnp") pod "4d11be72-00bb-4df8-91b4-8e8241ad618a" (UID: "4d11be72-00bb-4df8-91b4-8e8241ad618a"). InnerVolumeSpecName "kube-api-access-2zlnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.566094 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.566152 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zlnp\" (UniqueName: \"kubernetes.io/projected/4d11be72-00bb-4df8-91b4-8e8241ad618a-kube-api-access-2zlnp\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.599670 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d11be72-00bb-4df8-91b4-8e8241ad618a" (UID: "4d11be72-00bb-4df8-91b4-8e8241ad618a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.667227 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d11be72-00bb-4df8-91b4-8e8241ad618a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:26 crc kubenswrapper[4744]: I1201 08:33:26.866082 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52vz5"] Dec 01 08:33:27 crc kubenswrapper[4744]: I1201 08:33:27.306448 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcgv9" Dec 01 08:33:27 crc kubenswrapper[4744]: I1201 08:33:27.325350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcgv9" event={"ID":"4d11be72-00bb-4df8-91b4-8e8241ad618a","Type":"ContainerDied","Data":"8b8661aab4d1f072caff12c8f7b41c792f2387ab0a5fe93641f4d0cc3af60910"} Dec 01 08:33:27 crc kubenswrapper[4744]: I1201 08:33:27.325467 4744 scope.go:117] "RemoveContainer" containerID="04173ff773c21855372597767b524f31c269d1e4da36192d9582dc60baeb24bf" Dec 01 08:33:27 crc kubenswrapper[4744]: I1201 08:33:27.350341 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcgv9"] Dec 01 08:33:27 crc kubenswrapper[4744]: I1201 08:33:27.357395 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qcgv9"] Dec 01 08:33:27 crc kubenswrapper[4744]: I1201 08:33:27.357967 4744 scope.go:117] "RemoveContainer" containerID="abdc9fffc03f19b101be95fa2347579c069f0000a4be4f37a3d50d6568bb6611" Dec 01 08:33:27 crc kubenswrapper[4744]: I1201 08:33:27.462551 4744 scope.go:117] "RemoveContainer" containerID="80655c12bbba246396323ca0e90f22e1e9aa3576572d5515abba3d99cdc87d25" Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.312245 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" path="/var/lib/kubelet/pods/4d11be72-00bb-4df8-91b4-8e8241ad618a/volumes" Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.321575 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-52vz5" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="registry-server" containerID="cri-o://eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5" gracePeriod=2 Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.751741 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.899341 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-utilities\") pod \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.900001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gn6t\" (UniqueName: \"kubernetes.io/projected/f41f9cdf-d105-427a-a4e7-fb1f2320111d-kube-api-access-5gn6t\") pod \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.900035 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-catalog-content\") pod \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\" (UID: \"f41f9cdf-d105-427a-a4e7-fb1f2320111d\") " Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.901809 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-utilities" (OuterVolumeSpecName: "utilities") pod "f41f9cdf-d105-427a-a4e7-fb1f2320111d" (UID: "f41f9cdf-d105-427a-a4e7-fb1f2320111d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.906053 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f41f9cdf-d105-427a-a4e7-fb1f2320111d-kube-api-access-5gn6t" (OuterVolumeSpecName: "kube-api-access-5gn6t") pod "f41f9cdf-d105-427a-a4e7-fb1f2320111d" (UID: "f41f9cdf-d105-427a-a4e7-fb1f2320111d"). InnerVolumeSpecName "kube-api-access-5gn6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:28 crc kubenswrapper[4744]: I1201 08:33:28.933896 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f41f9cdf-d105-427a-a4e7-fb1f2320111d" (UID: "f41f9cdf-d105-427a-a4e7-fb1f2320111d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.001828 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.001880 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gn6t\" (UniqueName: \"kubernetes.io/projected/f41f9cdf-d105-427a-a4e7-fb1f2320111d-kube-api-access-5gn6t\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.001901 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41f9cdf-d105-427a-a4e7-fb1f2320111d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.337264 4744 generic.go:334] "Generic (PLEG): container finished" podID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerID="eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5" exitCode=0 Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.337333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52vz5" event={"ID":"f41f9cdf-d105-427a-a4e7-fb1f2320111d","Type":"ContainerDied","Data":"eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5"} Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.337468 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52vz5" event={"ID":"f41f9cdf-d105-427a-a4e7-fb1f2320111d","Type":"ContainerDied","Data":"a9dd8cc27fcca5641824897e0ab3c3b3786c73a0d3b064de3eb5d1ab3a262a29"} Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.337509 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52vz5" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.337579 4744 scope.go:117] "RemoveContainer" containerID="eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.382678 4744 scope.go:117] "RemoveContainer" containerID="1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.382689 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52vz5"] Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.389353 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-52vz5"] Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.426765 4744 scope.go:117] "RemoveContainer" containerID="87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.454759 4744 scope.go:117] "RemoveContainer" containerID="eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5" Dec 01 08:33:29 crc kubenswrapper[4744]: E1201 08:33:29.455555 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5\": container with ID starting with eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5 not found: ID does not exist" containerID="eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.455630 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5"} err="failed to get container status \"eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5\": rpc error: code = NotFound desc = could not find container \"eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5\": container with ID starting with eba9eec5a9aa07ccf0eb74491616ba0a62c09b85b877c25868abb524f8baeeb5 not found: ID does not exist" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.455684 4744 scope.go:117] "RemoveContainer" containerID="1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2" Dec 01 08:33:29 crc kubenswrapper[4744]: E1201 08:33:29.456157 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2\": container with ID starting with 1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2 not found: ID does not exist" containerID="1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.456257 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2"} err="failed to get container status \"1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2\": rpc error: code = NotFound desc = could not find container \"1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2\": container with ID starting with 1de6aac13046541a422ddc59c22be7884d56e9dfdcc4a681a8564bb7800457a2 not found: ID does not exist" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.456274 4744 scope.go:117] "RemoveContainer" containerID="87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d" Dec 01 08:33:29 crc kubenswrapper[4744]: E1201 08:33:29.457225 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d\": container with ID starting with 87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d not found: ID does not exist" containerID="87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d" Dec 01 08:33:29 crc kubenswrapper[4744]: I1201 08:33:29.457256 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d"} err="failed to get container status \"87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d\": rpc error: code = NotFound desc = could not find container \"87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d\": container with ID starting with 87f7317759053e92a9e16ec9f30d5b8875d59214b135d7de948fdc588851ef8d not found: ID does not exist" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.297541 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" path="/var/lib/kubelet/pods/f41f9cdf-d105-427a-a4e7-fb1f2320111d/volumes" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.378694 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w9nv6"] Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379226 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379242 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379270 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="extract-utilities" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379279 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="extract-utilities" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379300 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="extract-utilities" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379310 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="extract-utilities" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379323 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379331 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379340 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379348 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379361 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="extract-content" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379368 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="extract-content" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379395 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="extract-content" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379418 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="extract-content" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379432 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="extract-content" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379439 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="extract-content" Dec 01 08:33:30 crc kubenswrapper[4744]: E1201 08:33:30.379456 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="extract-utilities" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379473 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="extract-utilities" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379637 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d11be72-00bb-4df8-91b4-8e8241ad618a" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379647 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f0f866-0e29-42b0-8637-a7eab12cd951" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.379659 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41f9cdf-d105-427a-a4e7-fb1f2320111d" containerName="registry-server" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.380719 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.383080 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.383158 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8rdf9" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.383349 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.384164 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.403633 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w9nv6"] Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.458556 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z8vcb"] Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.468187 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.470601 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.472841 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z8vcb"] Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.543979 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltkwf\" (UniqueName: \"kubernetes.io/projected/23b7fb24-49ac-426b-b032-7e925abce5fa-kube-api-access-ltkwf\") pod \"dnsmasq-dns-675f4bcbfc-w9nv6\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.544032 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b7fb24-49ac-426b-b032-7e925abce5fa-config\") pod \"dnsmasq-dns-675f4bcbfc-w9nv6\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.544078 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-config\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.544099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.544140 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24nqg\" (UniqueName: \"kubernetes.io/projected/4d685b83-fb3e-4898-8e84-0b4b21c53db9-kube-api-access-24nqg\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.644989 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-config\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.645049 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.645088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24nqg\" (UniqueName: \"kubernetes.io/projected/4d685b83-fb3e-4898-8e84-0b4b21c53db9-kube-api-access-24nqg\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.645139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltkwf\" (UniqueName: \"kubernetes.io/projected/23b7fb24-49ac-426b-b032-7e925abce5fa-kube-api-access-ltkwf\") pod \"dnsmasq-dns-675f4bcbfc-w9nv6\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.645184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b7fb24-49ac-426b-b032-7e925abce5fa-config\") pod \"dnsmasq-dns-675f4bcbfc-w9nv6\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.646171 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b7fb24-49ac-426b-b032-7e925abce5fa-config\") pod \"dnsmasq-dns-675f4bcbfc-w9nv6\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.646175 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.646814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-config\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.688080 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24nqg\" (UniqueName: \"kubernetes.io/projected/4d685b83-fb3e-4898-8e84-0b4b21c53db9-kube-api-access-24nqg\") pod \"dnsmasq-dns-78dd6ddcc-z8vcb\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.688097 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltkwf\" (UniqueName: \"kubernetes.io/projected/23b7fb24-49ac-426b-b032-7e925abce5fa-kube-api-access-ltkwf\") pod \"dnsmasq-dns-675f4bcbfc-w9nv6\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.703113 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:30 crc kubenswrapper[4744]: I1201 08:33:30.785357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:31 crc kubenswrapper[4744]: I1201 08:33:31.192775 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z8vcb"] Dec 01 08:33:31 crc kubenswrapper[4744]: I1201 08:33:31.289025 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w9nv6"] Dec 01 08:33:31 crc kubenswrapper[4744]: I1201 08:33:31.363975 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" event={"ID":"4d685b83-fb3e-4898-8e84-0b4b21c53db9","Type":"ContainerStarted","Data":"598f7966e4857abcd62833c8b41f007b89358477696cb30a71260289afed6341"} Dec 01 08:33:31 crc kubenswrapper[4744]: I1201 08:33:31.365747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" event={"ID":"23b7fb24-49ac-426b-b032-7e925abce5fa","Type":"ContainerStarted","Data":"67c2018d645c345944c7b5fc7e1c3aa0bb301894cecd226ab5e75bcdd175f658"} Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.644894 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w9nv6"] Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.675290 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kzbvl"] Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.676646 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.694659 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kzbvl"] Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.794324 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqf29\" (UniqueName: \"kubernetes.io/projected/0e65135f-0286-4460-ae69-e3e9fb11b53a-kube-api-access-nqf29\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.794500 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.794567 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-config\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.896848 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqf29\" (UniqueName: \"kubernetes.io/projected/0e65135f-0286-4460-ae69-e3e9fb11b53a-kube-api-access-nqf29\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.897327 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.897358 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-config\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.898167 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.898192 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-config\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.921311 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqf29\" (UniqueName: \"kubernetes.io/projected/0e65135f-0286-4460-ae69-e3e9fb11b53a-kube-api-access-nqf29\") pod \"dnsmasq-dns-666b6646f7-kzbvl\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.972086 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z8vcb"] Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.994304 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6rw87"] Dec 01 08:33:33 crc kubenswrapper[4744]: I1201 08:33:33.995699 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.002274 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.006096 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6rw87"] Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.100086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqf24\" (UniqueName: \"kubernetes.io/projected/4fb767c7-9ffa-489e-a594-433e4cb2675e-kube-api-access-lqf24\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.100136 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.100264 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-config\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.201235 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-config\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.201305 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqf24\" (UniqueName: \"kubernetes.io/projected/4fb767c7-9ffa-489e-a594-433e4cb2675e-kube-api-access-lqf24\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.201339 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.202232 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.206803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-config\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.226708 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqf24\" (UniqueName: \"kubernetes.io/projected/4fb767c7-9ffa-489e-a594-433e4cb2675e-kube-api-access-lqf24\") pod \"dnsmasq-dns-57d769cc4f-6rw87\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.321658 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.561379 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kzbvl"] Dec 01 08:33:34 crc kubenswrapper[4744]: W1201 08:33:34.568943 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e65135f_0286_4460_ae69_e3e9fb11b53a.slice/crio-403c374edb413f37a397db2034e3e94ae92bc3f3dfcf6da5d541f03dc7c129c8 WatchSource:0}: Error finding container 403c374edb413f37a397db2034e3e94ae92bc3f3dfcf6da5d541f03dc7c129c8: Status 404 returned error can't find the container with id 403c374edb413f37a397db2034e3e94ae92bc3f3dfcf6da5d541f03dc7c129c8 Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.798575 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6rw87"] Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.816578 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.818231 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.820674 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.820821 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.820935 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.821076 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.821259 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.821585 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.821751 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x8gr9" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.828211 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918379 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf94bcff-f22a-4a35-beca-0096f08ee1f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918498 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-config-data\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918573 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf94bcff-f22a-4a35-beca-0096f08ee1f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918591 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h58q\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-kube-api-access-5h58q\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918620 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918706 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918813 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:34 crc kubenswrapper[4744]: I1201 08:33:34.918923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020364 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf94bcff-f22a-4a35-beca-0096f08ee1f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020439 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020469 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020488 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-config-data\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020503 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf94bcff-f22a-4a35-beca-0096f08ee1f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020535 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h58q\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-kube-api-access-5h58q\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020565 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020589 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.020621 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.021219 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.021474 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.021487 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.023677 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.023946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-config-data\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.025327 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.026190 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf94bcff-f22a-4a35-beca-0096f08ee1f6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.026715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.027231 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf94bcff-f22a-4a35-beca-0096f08ee1f6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.041132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.041965 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h58q\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-kube-api-access-5h58q\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.046528 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.100770 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.103825 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.110162 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.110357 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.110468 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.110596 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.110770 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.110888 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.110352 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-75bjt" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.131227 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.144167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.234419 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.234475 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/788a0c2b-cd13-48cb-99af-ba838e154525-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.234830 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.234865 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.245482 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtkgl\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-kube-api-access-dtkgl\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.245546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.245608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.245671 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.245717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/788a0c2b-cd13-48cb-99af-ba838e154525-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.245767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.245808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.347917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.347984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/788a0c2b-cd13-48cb-99af-ba838e154525-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.348012 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.348592 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.348628 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.348644 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/788a0c2b-cd13-48cb-99af-ba838e154525-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.348832 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.348862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.349027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.349400 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.349464 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.349609 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtkgl\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-kube-api-access-dtkgl\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.349650 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.349700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.352181 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.353274 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.363678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.363883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/788a0c2b-cd13-48cb-99af-ba838e154525-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.366137 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.375073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.377194 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/788a0c2b-cd13-48cb-99af-ba838e154525-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.379651 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtkgl\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-kube-api-access-dtkgl\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.380897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.406490 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" event={"ID":"0e65135f-0286-4460-ae69-e3e9fb11b53a","Type":"ContainerStarted","Data":"403c374edb413f37a397db2034e3e94ae92bc3f3dfcf6da5d541f03dc7c129c8"} Dec 01 08:33:35 crc kubenswrapper[4744]: I1201 08:33:35.443770 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.453392 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.455061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.462449 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.463070 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-w62rg" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.463111 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.463701 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.467791 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.479988 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.568885 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-kolla-config\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.568944 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mfb6\" (UniqueName: \"kubernetes.io/projected/0b711f28-53bd-46e8-9a86-44be43bd78d3-kube-api-access-6mfb6\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.568970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-config-data-default\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.568995 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.569062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b711f28-53bd-46e8-9a86-44be43bd78d3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.569315 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b711f28-53bd-46e8-9a86-44be43bd78d3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.569470 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.569557 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b711f28-53bd-46e8-9a86-44be43bd78d3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.671930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b711f28-53bd-46e8-9a86-44be43bd78d3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.672019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.672069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b711f28-53bd-46e8-9a86-44be43bd78d3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.672101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-kolla-config\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.672128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mfb6\" (UniqueName: \"kubernetes.io/projected/0b711f28-53bd-46e8-9a86-44be43bd78d3-kube-api-access-6mfb6\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.672160 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-config-data-default\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.672186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.672235 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b711f28-53bd-46e8-9a86-44be43bd78d3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.673716 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.673783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-config-data-default\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.681084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b711f28-53bd-46e8-9a86-44be43bd78d3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.681925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.683077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b711f28-53bd-46e8-9a86-44be43bd78d3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.688083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b711f28-53bd-46e8-9a86-44be43bd78d3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.692051 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mfb6\" (UniqueName: \"kubernetes.io/projected/0b711f28-53bd-46e8-9a86-44be43bd78d3-kube-api-access-6mfb6\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.693034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b711f28-53bd-46e8-9a86-44be43bd78d3-kolla-config\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.698652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0b711f28-53bd-46e8-9a86-44be43bd78d3\") " pod="openstack/openstack-galera-0" Dec 01 08:33:36 crc kubenswrapper[4744]: I1201 08:33:36.774309 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 08:33:37 crc kubenswrapper[4744]: I1201 08:33:37.848975 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 08:33:37 crc kubenswrapper[4744]: I1201 08:33:37.850366 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:37 crc kubenswrapper[4744]: I1201 08:33:37.852106 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 08:33:37 crc kubenswrapper[4744]: I1201 08:33:37.853031 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 08:33:37 crc kubenswrapper[4744]: I1201 08:33:37.853131 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 08:33:37 crc kubenswrapper[4744]: I1201 08:33:37.855061 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-whp6s" Dec 01 08:33:37 crc kubenswrapper[4744]: I1201 08:33:37.877875 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007260 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31db90c-376f-45b5-9d13-73bcf4962135-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007368 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42xmx\" (UniqueName: \"kubernetes.io/projected/c31db90c-376f-45b5-9d13-73bcf4962135-kube-api-access-42xmx\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c31db90c-376f-45b5-9d13-73bcf4962135-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31db90c-376f-45b5-9d13-73bcf4962135-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007501 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.007549 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.109465 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.109624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31db90c-376f-45b5-9d13-73bcf4962135-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.109873 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42xmx\" (UniqueName: \"kubernetes.io/projected/c31db90c-376f-45b5-9d13-73bcf4962135-kube-api-access-42xmx\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.110731 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.110784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.111101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c31db90c-376f-45b5-9d13-73bcf4962135-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.111545 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c31db90c-376f-45b5-9d13-73bcf4962135-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.111610 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31db90c-376f-45b5-9d13-73bcf4962135-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.111691 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.111941 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.112203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.112323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.115073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31db90c-376f-45b5-9d13-73bcf4962135-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.115720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31db90c-376f-45b5-9d13-73bcf4962135-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.118226 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c31db90c-376f-45b5-9d13-73bcf4962135-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.137381 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42xmx\" (UniqueName: \"kubernetes.io/projected/c31db90c-376f-45b5-9d13-73bcf4962135-kube-api-access-42xmx\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.140490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c31db90c-376f-45b5-9d13-73bcf4962135\") " pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.170513 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.432624 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.433687 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.439256 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.439590 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hg4st" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.439762 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.487472 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.526530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0153c296-2e70-43fa-90ae-bb007406d2c0-config-data\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.526589 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0153c296-2e70-43fa-90ae-bb007406d2c0-kolla-config\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.526624 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0153c296-2e70-43fa-90ae-bb007406d2c0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.526640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0153c296-2e70-43fa-90ae-bb007406d2c0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.526661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrqsj\" (UniqueName: \"kubernetes.io/projected/0153c296-2e70-43fa-90ae-bb007406d2c0-kube-api-access-mrqsj\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.627518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0153c296-2e70-43fa-90ae-bb007406d2c0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.627572 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0153c296-2e70-43fa-90ae-bb007406d2c0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.627599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrqsj\" (UniqueName: \"kubernetes.io/projected/0153c296-2e70-43fa-90ae-bb007406d2c0-kube-api-access-mrqsj\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.627662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0153c296-2e70-43fa-90ae-bb007406d2c0-config-data\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.627694 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0153c296-2e70-43fa-90ae-bb007406d2c0-kolla-config\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.628349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0153c296-2e70-43fa-90ae-bb007406d2c0-kolla-config\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.629622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0153c296-2e70-43fa-90ae-bb007406d2c0-config-data\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.632790 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0153c296-2e70-43fa-90ae-bb007406d2c0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.633392 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0153c296-2e70-43fa-90ae-bb007406d2c0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.645848 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrqsj\" (UniqueName: \"kubernetes.io/projected/0153c296-2e70-43fa-90ae-bb007406d2c0-kube-api-access-mrqsj\") pod \"memcached-0\" (UID: \"0153c296-2e70-43fa-90ae-bb007406d2c0\") " pod="openstack/memcached-0" Dec 01 08:33:38 crc kubenswrapper[4744]: I1201 08:33:38.761220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.068967 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.070251 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.075297 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bm8l9" Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.081043 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.153261 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4mrh\" (UniqueName: \"kubernetes.io/projected/e70a50a9-9bb9-45be-aa60-c57c67e9fc32-kube-api-access-z4mrh\") pod \"kube-state-metrics-0\" (UID: \"e70a50a9-9bb9-45be-aa60-c57c67e9fc32\") " pod="openstack/kube-state-metrics-0" Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.254206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4mrh\" (UniqueName: \"kubernetes.io/projected/e70a50a9-9bb9-45be-aa60-c57c67e9fc32-kube-api-access-z4mrh\") pod \"kube-state-metrics-0\" (UID: \"e70a50a9-9bb9-45be-aa60-c57c67e9fc32\") " pod="openstack/kube-state-metrics-0" Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.272323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4mrh\" (UniqueName: \"kubernetes.io/projected/e70a50a9-9bb9-45be-aa60-c57c67e9fc32-kube-api-access-z4mrh\") pod \"kube-state-metrics-0\" (UID: \"e70a50a9-9bb9-45be-aa60-c57c67e9fc32\") " pod="openstack/kube-state-metrics-0" Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.394506 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:33:40 crc kubenswrapper[4744]: I1201 08:33:40.447467 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" event={"ID":"4fb767c7-9ffa-489e-a594-433e4cb2675e","Type":"ContainerStarted","Data":"e37392125d67ffb56a82e2e8be4467040c6a54d2a1f7ce5066b1220dcf68352d"} Dec 01 08:33:42 crc kubenswrapper[4744]: I1201 08:33:42.575088 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.181955 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pgp2f"] Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.183511 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.185512 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-48gsf" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.185725 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.191516 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pgp2f"] Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.192020 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.201119 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-fzxwj"] Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.204761 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.212432 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fzxwj"] Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.311964 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nffxv\" (UniqueName: \"kubernetes.io/projected/ac7e9e72-1bca-42aa-a062-e9743651f2aa-kube-api-access-nffxv\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-run\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312068 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7e9e72-1bca-42aa-a062-e9743651f2aa-ovn-controller-tls-certs\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312171 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac7e9e72-1bca-42aa-a062-e9743651f2aa-scripts\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312284 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-log-ovn\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312348 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b14f48b-6456-4329-97fa-e81ec05e2644-scripts\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312372 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqmdf\" (UniqueName: \"kubernetes.io/projected/2b14f48b-6456-4329-97fa-e81ec05e2644-kube-api-access-dqmdf\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312493 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-lib\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312573 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7e9e72-1bca-42aa-a062-e9743651f2aa-combined-ca-bundle\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312649 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-run-ovn\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-run\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312721 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-etc-ovs\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.312749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-log\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.413929 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-log-ovn\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.413995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqmdf\" (UniqueName: \"kubernetes.io/projected/2b14f48b-6456-4329-97fa-e81ec05e2644-kube-api-access-dqmdf\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414021 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b14f48b-6456-4329-97fa-e81ec05e2644-scripts\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-lib\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414123 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7e9e72-1bca-42aa-a062-e9743651f2aa-combined-ca-bundle\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414164 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-run\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414182 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-run-ovn\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414208 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-etc-ovs\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414245 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-log\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414267 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nffxv\" (UniqueName: \"kubernetes.io/projected/ac7e9e72-1bca-42aa-a062-e9743651f2aa-kube-api-access-nffxv\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-run\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7e9e72-1bca-42aa-a062-e9743651f2aa-ovn-controller-tls-certs\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414350 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac7e9e72-1bca-42aa-a062-e9743651f2aa-scripts\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414572 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-log-ovn\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-run\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414716 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-run\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414839 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ac7e9e72-1bca-42aa-a062-e9743651f2aa-var-run-ovn\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.414894 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-etc-ovs\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.415075 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-lib\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.415240 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2b14f48b-6456-4329-97fa-e81ec05e2644-var-log\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.418180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b14f48b-6456-4329-97fa-e81ec05e2644-scripts\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.418210 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac7e9e72-1bca-42aa-a062-e9743651f2aa-scripts\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.420194 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7e9e72-1bca-42aa-a062-e9743651f2aa-combined-ca-bundle\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.426774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7e9e72-1bca-42aa-a062-e9743651f2aa-ovn-controller-tls-certs\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.432543 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nffxv\" (UniqueName: \"kubernetes.io/projected/ac7e9e72-1bca-42aa-a062-e9743651f2aa-kube-api-access-nffxv\") pod \"ovn-controller-pgp2f\" (UID: \"ac7e9e72-1bca-42aa-a062-e9743651f2aa\") " pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.447186 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqmdf\" (UniqueName: \"kubernetes.io/projected/2b14f48b-6456-4329-97fa-e81ec05e2644-kube-api-access-dqmdf\") pod \"ovn-controller-ovs-fzxwj\" (UID: \"2b14f48b-6456-4329-97fa-e81ec05e2644\") " pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.513796 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pgp2f" Dec 01 08:33:44 crc kubenswrapper[4744]: I1201 08:33:44.522994 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:33:46 crc kubenswrapper[4744]: W1201 08:33:46.798592 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod788a0c2b_cd13_48cb_99af_ba838e154525.slice/crio-7cff00ef7426c20912aca14f985a667dd253ac736be714672885555d071f5104 WatchSource:0}: Error finding container 7cff00ef7426c20912aca14f985a667dd253ac736be714672885555d071f5104: Status 404 returned error can't find the container with id 7cff00ef7426c20912aca14f985a667dd253ac736be714672885555d071f5104 Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.882769 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.884889 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.889215 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.889499 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rh6kv" Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.889720 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.890378 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.890623 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 08:33:46 crc kubenswrapper[4744]: I1201 08:33:46.904890 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.052893 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.053230 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.053260 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bjf2\" (UniqueName: \"kubernetes.io/projected/a6ebb51f-ba73-417a-b6a8-ff31ff555108-kube-api-access-4bjf2\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.053282 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.053331 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6ebb51f-ba73-417a-b6a8-ff31ff555108-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.053358 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.053382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6ebb51f-ba73-417a-b6a8-ff31ff555108-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.053420 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6ebb51f-ba73-417a-b6a8-ff31ff555108-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.084660 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.086137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.090238 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.090268 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.090290 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.090318 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-g9dcn" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.095462 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bjf2\" (UniqueName: \"kubernetes.io/projected/a6ebb51f-ba73-417a-b6a8-ff31ff555108-kube-api-access-4bjf2\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155252 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6ebb51f-ba73-417a-b6a8-ff31ff555108-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155328 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155345 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6ebb51f-ba73-417a-b6a8-ff31ff555108-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155363 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6ebb51f-ba73-417a-b6a8-ff31ff555108-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155625 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.155827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6ebb51f-ba73-417a-b6a8-ff31ff555108-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.156258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6ebb51f-ba73-417a-b6a8-ff31ff555108-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.156671 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6ebb51f-ba73-417a-b6a8-ff31ff555108-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.162062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.162680 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.164066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ebb51f-ba73-417a-b6a8-ff31ff555108-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.179230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bjf2\" (UniqueName: \"kubernetes.io/projected/a6ebb51f-ba73-417a-b6a8-ff31ff555108-kube-api-access-4bjf2\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.179916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6ebb51f-ba73-417a-b6a8-ff31ff555108\") " pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.218757 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.227273 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256339 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256427 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bptxk\" (UniqueName: \"kubernetes.io/projected/916bb441-60a0-4452-b2cf-5b98fb9aea9f-kube-api-access-bptxk\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256492 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256700 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916bb441-60a0-4452-b2cf-5b98fb9aea9f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256840 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916bb441-60a0-4452-b2cf-5b98fb9aea9f-config\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256886 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/916bb441-60a0-4452-b2cf-5b98fb9aea9f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.256930 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.309824 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.360889 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.360985 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916bb441-60a0-4452-b2cf-5b98fb9aea9f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.361021 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916bb441-60a0-4452-b2cf-5b98fb9aea9f-config\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.361054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/916bb441-60a0-4452-b2cf-5b98fb9aea9f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.361093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.361121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.361135 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.361150 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.361158 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bptxk\" (UniqueName: \"kubernetes.io/projected/916bb441-60a0-4452-b2cf-5b98fb9aea9f-kube-api-access-bptxk\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.363397 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916bb441-60a0-4452-b2cf-5b98fb9aea9f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.363945 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916bb441-60a0-4452-b2cf-5b98fb9aea9f-config\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.365072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/916bb441-60a0-4452-b2cf-5b98fb9aea9f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.376498 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.377006 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.377180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/916bb441-60a0-4452-b2cf-5b98fb9aea9f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.381757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bptxk\" (UniqueName: \"kubernetes.io/projected/916bb441-60a0-4452-b2cf-5b98fb9aea9f-kube-api-access-bptxk\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.390722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"916bb441-60a0-4452-b2cf-5b98fb9aea9f\") " pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.413873 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 08:33:47 crc kubenswrapper[4744]: I1201 08:33:47.498928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"788a0c2b-cd13-48cb-99af-ba838e154525","Type":"ContainerStarted","Data":"7cff00ef7426c20912aca14f985a667dd253ac736be714672885555d071f5104"} Dec 01 08:33:47 crc kubenswrapper[4744]: W1201 08:33:47.811424 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc31db90c_376f_45b5_9d13_73bcf4962135.slice/crio-3f59237150a22df2ba84d19c6beebef70f3d583ec16f310174c58a370af4cbed WatchSource:0}: Error finding container 3f59237150a22df2ba84d19c6beebef70f3d583ec16f310174c58a370af4cbed: Status 404 returned error can't find the container with id 3f59237150a22df2ba84d19c6beebef70f3d583ec16f310174c58a370af4cbed Dec 01 08:33:47 crc kubenswrapper[4744]: E1201 08:33:47.816395 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 08:33:47 crc kubenswrapper[4744]: E1201 08:33:47.816590 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ltkwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-w9nv6_openstack(23b7fb24-49ac-426b-b032-7e925abce5fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:33:47 crc kubenswrapper[4744]: E1201 08:33:47.817782 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" podUID="23b7fb24-49ac-426b-b032-7e925abce5fa" Dec 01 08:33:47 crc kubenswrapper[4744]: E1201 08:33:47.841328 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 08:33:47 crc kubenswrapper[4744]: E1201 08:33:47.841943 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-24nqg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-z8vcb_openstack(4d685b83-fb3e-4898-8e84-0b4b21c53db9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:33:47 crc kubenswrapper[4744]: E1201 08:33:47.843246 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" podUID="4d685b83-fb3e-4898-8e84-0b4b21c53db9" Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.272808 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.387944 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:33:48 crc kubenswrapper[4744]: W1201 08:33:48.393964 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode70a50a9_9bb9_45be_aa60_c57c67e9fc32.slice/crio-26ebc04757163c14cb46ef58235eeb675be3c384777e27f94144c670346f7d9c WatchSource:0}: Error finding container 26ebc04757163c14cb46ef58235eeb675be3c384777e27f94144c670346f7d9c: Status 404 returned error can't find the container with id 26ebc04757163c14cb46ef58235eeb675be3c384777e27f94144c670346f7d9c Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.395619 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.508593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0153c296-2e70-43fa-90ae-bb007406d2c0","Type":"ContainerStarted","Data":"fad99d32bff223e558d47014b7026bfdfa13d38dea0aab9c3fad08ae78ade253"} Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.510233 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c31db90c-376f-45b5-9d13-73bcf4962135","Type":"ContainerStarted","Data":"3f59237150a22df2ba84d19c6beebef70f3d583ec16f310174c58a370af4cbed"} Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.512024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b711f28-53bd-46e8-9a86-44be43bd78d3","Type":"ContainerStarted","Data":"b8dee6503928a3f56f502d55f809b7624b24af576b27ef5e4fac1934047e873d"} Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.513179 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bf94bcff-f22a-4a35-beca-0096f08ee1f6","Type":"ContainerStarted","Data":"f60789b4e08fa98fbc562d71cfeeefd3c88476b25244c6a2ed8278cbea2a0b9c"} Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.515543 4744 generic.go:334] "Generic (PLEG): container finished" podID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerID="5f1b8d0d04914a5912e725e99f62a62d4dec2df8dd5aba6413956122e9c539c7" exitCode=0 Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.515619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" event={"ID":"4fb767c7-9ffa-489e-a594-433e4cb2675e","Type":"ContainerDied","Data":"5f1b8d0d04914a5912e725e99f62a62d4dec2df8dd5aba6413956122e9c539c7"} Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.517968 4744 generic.go:334] "Generic (PLEG): container finished" podID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerID="5fedfc92859d2953ed3168e011a2cb520427590d0ba30c081b9f8efc81822e22" exitCode=0 Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.518039 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" event={"ID":"0e65135f-0286-4460-ae69-e3e9fb11b53a","Type":"ContainerDied","Data":"5fedfc92859d2953ed3168e011a2cb520427590d0ba30c081b9f8efc81822e22"} Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.519165 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e70a50a9-9bb9-45be-aa60-c57c67e9fc32","Type":"ContainerStarted","Data":"26ebc04757163c14cb46ef58235eeb675be3c384777e27f94144c670346f7d9c"} Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.615549 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fzxwj"] Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.636629 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pgp2f"] Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.637040 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.637086 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:33:48 crc kubenswrapper[4744]: W1201 08:33:48.639425 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b14f48b_6456_4329_97fa_e81ec05e2644.slice/crio-2abf7b270d3e99476b250ca4c2841f9fab15e79b84a0b99f2baf38b4a5f1d61b WatchSource:0}: Error finding container 2abf7b270d3e99476b250ca4c2841f9fab15e79b84a0b99f2baf38b4a5f1d61b: Status 404 returned error can't find the container with id 2abf7b270d3e99476b250ca4c2841f9fab15e79b84a0b99f2baf38b4a5f1d61b Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.699258 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.946612 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:48 crc kubenswrapper[4744]: I1201 08:33:48.956953 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.089945 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltkwf\" (UniqueName: \"kubernetes.io/projected/23b7fb24-49ac-426b-b032-7e925abce5fa-kube-api-access-ltkwf\") pod \"23b7fb24-49ac-426b-b032-7e925abce5fa\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.090019 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b7fb24-49ac-426b-b032-7e925abce5fa-config\") pod \"23b7fb24-49ac-426b-b032-7e925abce5fa\" (UID: \"23b7fb24-49ac-426b-b032-7e925abce5fa\") " Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.090101 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24nqg\" (UniqueName: \"kubernetes.io/projected/4d685b83-fb3e-4898-8e84-0b4b21c53db9-kube-api-access-24nqg\") pod \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.090135 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-dns-svc\") pod \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.090179 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-config\") pod \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\" (UID: \"4d685b83-fb3e-4898-8e84-0b4b21c53db9\") " Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.090926 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-config" (OuterVolumeSpecName: "config") pod "4d685b83-fb3e-4898-8e84-0b4b21c53db9" (UID: "4d685b83-fb3e-4898-8e84-0b4b21c53db9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.096288 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23b7fb24-49ac-426b-b032-7e925abce5fa-kube-api-access-ltkwf" (OuterVolumeSpecName: "kube-api-access-ltkwf") pod "23b7fb24-49ac-426b-b032-7e925abce5fa" (UID: "23b7fb24-49ac-426b-b032-7e925abce5fa"). InnerVolumeSpecName "kube-api-access-ltkwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.096427 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d685b83-fb3e-4898-8e84-0b4b21c53db9" (UID: "4d685b83-fb3e-4898-8e84-0b4b21c53db9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.096423 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23b7fb24-49ac-426b-b032-7e925abce5fa-config" (OuterVolumeSpecName: "config") pod "23b7fb24-49ac-426b-b032-7e925abce5fa" (UID: "23b7fb24-49ac-426b-b032-7e925abce5fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.097251 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d685b83-fb3e-4898-8e84-0b4b21c53db9-kube-api-access-24nqg" (OuterVolumeSpecName: "kube-api-access-24nqg") pod "4d685b83-fb3e-4898-8e84-0b4b21c53db9" (UID: "4d685b83-fb3e-4898-8e84-0b4b21c53db9"). InnerVolumeSpecName "kube-api-access-24nqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.192230 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.192267 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltkwf\" (UniqueName: \"kubernetes.io/projected/23b7fb24-49ac-426b-b032-7e925abce5fa-kube-api-access-ltkwf\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.192300 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b7fb24-49ac-426b-b032-7e925abce5fa-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.192312 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24nqg\" (UniqueName: \"kubernetes.io/projected/4d685b83-fb3e-4898-8e84-0b4b21c53db9-kube-api-access-24nqg\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.192321 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d685b83-fb3e-4898-8e84-0b4b21c53db9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.493723 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 08:33:49 crc kubenswrapper[4744]: W1201 08:33:49.514556 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6ebb51f_ba73_417a_b6a8_ff31ff555108.slice/crio-dbd554697d5b3bd1c6643207e63409a87938ff71c112b25d6459b4c855bc232f WatchSource:0}: Error finding container dbd554697d5b3bd1c6643207e63409a87938ff71c112b25d6459b4c855bc232f: Status 404 returned error can't find the container with id dbd554697d5b3bd1c6643207e63409a87938ff71c112b25d6459b4c855bc232f Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.530722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6ebb51f-ba73-417a-b6a8-ff31ff555108","Type":"ContainerStarted","Data":"dbd554697d5b3bd1c6643207e63409a87938ff71c112b25d6459b4c855bc232f"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.532333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fzxwj" event={"ID":"2b14f48b-6456-4329-97fa-e81ec05e2644","Type":"ContainerStarted","Data":"2abf7b270d3e99476b250ca4c2841f9fab15e79b84a0b99f2baf38b4a5f1d61b"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.534969 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" event={"ID":"4d685b83-fb3e-4898-8e84-0b4b21c53db9","Type":"ContainerDied","Data":"598f7966e4857abcd62833c8b41f007b89358477696cb30a71260289afed6341"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.535079 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z8vcb" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.541655 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"916bb441-60a0-4452-b2cf-5b98fb9aea9f","Type":"ContainerStarted","Data":"e378e08b856801e2ecd4083f24805e84bd4269e028e92d7c7b8b3b65394c3336"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.547471 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pgp2f" event={"ID":"ac7e9e72-1bca-42aa-a062-e9743651f2aa","Type":"ContainerStarted","Data":"5c050fe48f6d8a699352dc8858c783ac319e8fd587223ae94121a45a685a6693"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.554226 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" event={"ID":"4fb767c7-9ffa-489e-a594-433e4cb2675e","Type":"ContainerStarted","Data":"befbd6837b723fccd50c63367c4d50be1e3d1e7aa1b33893167f0156fa613e00"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.554324 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.557837 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.557846 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-w9nv6" event={"ID":"23b7fb24-49ac-426b-b032-7e925abce5fa","Type":"ContainerDied","Data":"67c2018d645c345944c7b5fc7e1c3aa0bb301894cecd226ab5e75bcdd175f658"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.560308 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" event={"ID":"0e65135f-0286-4460-ae69-e3e9fb11b53a","Type":"ContainerStarted","Data":"2527d9b8b6db2d7133930c1c33da958e7f29241febb320daa90a99ae4cc58481"} Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.560449 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.634390 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z8vcb"] Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.645888 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z8vcb"] Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.652465 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" podStartSLOduration=8.289458515 podStartE2EDuration="16.65243944s" podCreationTimestamp="2025-12-01 08:33:33 +0000 UTC" firstStartedPulling="2025-12-01 08:33:39.587768816 +0000 UTC m=+951.576826747" lastFinishedPulling="2025-12-01 08:33:47.950749751 +0000 UTC m=+959.939807672" observedRunningTime="2025-12-01 08:33:49.606783419 +0000 UTC m=+961.595841350" watchObservedRunningTime="2025-12-01 08:33:49.65243944 +0000 UTC m=+961.641497361" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.654957 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" podStartSLOduration=3.3010483600000002 podStartE2EDuration="16.654932102s" podCreationTimestamp="2025-12-01 08:33:33 +0000 UTC" firstStartedPulling="2025-12-01 08:33:34.57428686 +0000 UTC m=+946.563344781" lastFinishedPulling="2025-12-01 08:33:47.928170602 +0000 UTC m=+959.917228523" observedRunningTime="2025-12-01 08:33:49.643469282 +0000 UTC m=+961.632527203" watchObservedRunningTime="2025-12-01 08:33:49.654932102 +0000 UTC m=+961.643990023" Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.683253 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w9nv6"] Dec 01 08:33:49 crc kubenswrapper[4744]: I1201 08:33:49.688000 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w9nv6"] Dec 01 08:33:50 crc kubenswrapper[4744]: I1201 08:33:50.304828 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23b7fb24-49ac-426b-b032-7e925abce5fa" path="/var/lib/kubelet/pods/23b7fb24-49ac-426b-b032-7e925abce5fa/volumes" Dec 01 08:33:50 crc kubenswrapper[4744]: I1201 08:33:50.305168 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d685b83-fb3e-4898-8e84-0b4b21c53db9" path="/var/lib/kubelet/pods/4d685b83-fb3e-4898-8e84-0b4b21c53db9/volumes" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.113767 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-4nncr"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.115383 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.124804 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.131828 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4nncr"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.227117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efccc7e5-6075-43fb-b36e-15c4734e8f97-combined-ca-bundle\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.227176 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/efccc7e5-6075-43fb-b36e-15c4734e8f97-ovn-rundir\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.227200 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/efccc7e5-6075-43fb-b36e-15c4734e8f97-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.227225 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efccc7e5-6075-43fb-b36e-15c4734e8f97-config\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.227253 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hcgk\" (UniqueName: \"kubernetes.io/projected/efccc7e5-6075-43fb-b36e-15c4734e8f97-kube-api-access-5hcgk\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.227280 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/efccc7e5-6075-43fb-b36e-15c4734e8f97-ovs-rundir\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.269049 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6rw87"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.303226 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-59wd5"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.304550 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.306972 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.310790 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-59wd5"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.328817 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efccc7e5-6075-43fb-b36e-15c4734e8f97-combined-ca-bundle\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.328868 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/efccc7e5-6075-43fb-b36e-15c4734e8f97-ovn-rundir\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.328889 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/efccc7e5-6075-43fb-b36e-15c4734e8f97-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.328911 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efccc7e5-6075-43fb-b36e-15c4734e8f97-config\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.328938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hcgk\" (UniqueName: \"kubernetes.io/projected/efccc7e5-6075-43fb-b36e-15c4734e8f97-kube-api-access-5hcgk\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.328961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/efccc7e5-6075-43fb-b36e-15c4734e8f97-ovs-rundir\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.329261 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/efccc7e5-6075-43fb-b36e-15c4734e8f97-ovs-rundir\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.329897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/efccc7e5-6075-43fb-b36e-15c4734e8f97-ovn-rundir\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.331807 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efccc7e5-6075-43fb-b36e-15c4734e8f97-config\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.338317 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efccc7e5-6075-43fb-b36e-15c4734e8f97-combined-ca-bundle\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.338854 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/efccc7e5-6075-43fb-b36e-15c4734e8f97-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.346885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hcgk\" (UniqueName: \"kubernetes.io/projected/efccc7e5-6075-43fb-b36e-15c4734e8f97-kube-api-access-5hcgk\") pod \"ovn-controller-metrics-4nncr\" (UID: \"efccc7e5-6075-43fb-b36e-15c4734e8f97\") " pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.430736 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.430814 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.430849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcjqw\" (UniqueName: \"kubernetes.io/projected/403a172d-1da3-49fe-b553-5f930ec31720-kube-api-access-vcjqw\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.430895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-config\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.440643 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4nncr" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.516555 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kzbvl"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.533997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.534085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.534109 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcjqw\" (UniqueName: \"kubernetes.io/projected/403a172d-1da3-49fe-b553-5f930ec31720-kube-api-access-vcjqw\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.534145 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-config\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.535675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-config\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.536495 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.537059 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.552957 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-v49zx"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.556271 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.558476 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.568446 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-v49zx"] Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.569856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcjqw\" (UniqueName: \"kubernetes.io/projected/403a172d-1da3-49fe-b553-5f930ec31720-kube-api-access-vcjqw\") pod \"dnsmasq-dns-6bc7876d45-59wd5\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.605710 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" podUID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerName="dnsmasq-dns" containerID="cri-o://befbd6837b723fccd50c63367c4d50be1e3d1e7aa1b33893167f0156fa613e00" gracePeriod=10 Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.606617 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e70a50a9-9bb9-45be-aa60-c57c67e9fc32","Type":"ContainerStarted","Data":"089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680"} Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.606649 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.606745 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" podUID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerName="dnsmasq-dns" containerID="cri-o://2527d9b8b6db2d7133930c1c33da958e7f29241febb320daa90a99ae4cc58481" gracePeriod=10 Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.625098 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.473330846 podStartE2EDuration="11.625079099s" podCreationTimestamp="2025-12-01 08:33:40 +0000 UTC" firstStartedPulling="2025-12-01 08:33:48.396110471 +0000 UTC m=+960.385168392" lastFinishedPulling="2025-12-01 08:33:50.547858724 +0000 UTC m=+962.536916645" observedRunningTime="2025-12-01 08:33:51.624559964 +0000 UTC m=+963.613617885" watchObservedRunningTime="2025-12-01 08:33:51.625079099 +0000 UTC m=+963.614137010" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.628593 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.635211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clzxr\" (UniqueName: \"kubernetes.io/projected/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-kube-api-access-clzxr\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.635421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.635507 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-config\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.635594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.635690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-dns-svc\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.736666 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clzxr\" (UniqueName: \"kubernetes.io/projected/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-kube-api-access-clzxr\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.736762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.736798 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-config\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.736823 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.736863 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-dns-svc\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.738150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-config\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.738572 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.738592 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.738696 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-dns-svc\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.757833 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clzxr\" (UniqueName: \"kubernetes.io/projected/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-kube-api-access-clzxr\") pod \"dnsmasq-dns-8554648995-v49zx\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:51 crc kubenswrapper[4744]: I1201 08:33:51.907732 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:33:52 crc kubenswrapper[4744]: I1201 08:33:52.022326 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4nncr"] Dec 01 08:33:52 crc kubenswrapper[4744]: I1201 08:33:52.353573 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-59wd5"] Dec 01 08:33:52 crc kubenswrapper[4744]: I1201 08:33:52.619747 4744 generic.go:334] "Generic (PLEG): container finished" podID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerID="befbd6837b723fccd50c63367c4d50be1e3d1e7aa1b33893167f0156fa613e00" exitCode=0 Dec 01 08:33:52 crc kubenswrapper[4744]: I1201 08:33:52.619807 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" event={"ID":"4fb767c7-9ffa-489e-a594-433e4cb2675e","Type":"ContainerDied","Data":"befbd6837b723fccd50c63367c4d50be1e3d1e7aa1b33893167f0156fa613e00"} Dec 01 08:33:52 crc kubenswrapper[4744]: I1201 08:33:52.623326 4744 generic.go:334] "Generic (PLEG): container finished" podID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerID="2527d9b8b6db2d7133930c1c33da958e7f29241febb320daa90a99ae4cc58481" exitCode=0 Dec 01 08:33:52 crc kubenswrapper[4744]: I1201 08:33:52.623377 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" event={"ID":"0e65135f-0286-4460-ae69-e3e9fb11b53a","Type":"ContainerDied","Data":"2527d9b8b6db2d7133930c1c33da958e7f29241febb320daa90a99ae4cc58481"} Dec 01 08:33:53 crc kubenswrapper[4744]: W1201 08:33:53.160922 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod403a172d_1da3_49fe_b553_5f930ec31720.slice/crio-fa39b7a54b30530eaf7d5d6b7050738185bd04f53ed8bf2e1c71a4cd14c769be WatchSource:0}: Error finding container fa39b7a54b30530eaf7d5d6b7050738185bd04f53ed8bf2e1c71a4cd14c769be: Status 404 returned error can't find the container with id fa39b7a54b30530eaf7d5d6b7050738185bd04f53ed8bf2e1c71a4cd14c769be Dec 01 08:33:53 crc kubenswrapper[4744]: W1201 08:33:53.177945 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefccc7e5_6075_43fb_b36e_15c4734e8f97.slice/crio-7fab1bf3db64ec3d5bdf0a3f8c13ed5600e5ade65ae31c01df58aec6ab918602 WatchSource:0}: Error finding container 7fab1bf3db64ec3d5bdf0a3f8c13ed5600e5ade65ae31c01df58aec6ab918602: Status 404 returned error can't find the container with id 7fab1bf3db64ec3d5bdf0a3f8c13ed5600e5ade65ae31c01df58aec6ab918602 Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.230767 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.237152 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.383356 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-dns-svc\") pod \"0e65135f-0286-4460-ae69-e3e9fb11b53a\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.383393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-config\") pod \"4fb767c7-9ffa-489e-a594-433e4cb2675e\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.383454 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-config\") pod \"0e65135f-0286-4460-ae69-e3e9fb11b53a\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.383472 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqf29\" (UniqueName: \"kubernetes.io/projected/0e65135f-0286-4460-ae69-e3e9fb11b53a-kube-api-access-nqf29\") pod \"0e65135f-0286-4460-ae69-e3e9fb11b53a\" (UID: \"0e65135f-0286-4460-ae69-e3e9fb11b53a\") " Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.383574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqf24\" (UniqueName: \"kubernetes.io/projected/4fb767c7-9ffa-489e-a594-433e4cb2675e-kube-api-access-lqf24\") pod \"4fb767c7-9ffa-489e-a594-433e4cb2675e\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.383595 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-dns-svc\") pod \"4fb767c7-9ffa-489e-a594-433e4cb2675e\" (UID: \"4fb767c7-9ffa-489e-a594-433e4cb2675e\") " Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.388172 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fb767c7-9ffa-489e-a594-433e4cb2675e-kube-api-access-lqf24" (OuterVolumeSpecName: "kube-api-access-lqf24") pod "4fb767c7-9ffa-489e-a594-433e4cb2675e" (UID: "4fb767c7-9ffa-489e-a594-433e4cb2675e"). InnerVolumeSpecName "kube-api-access-lqf24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.389015 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e65135f-0286-4460-ae69-e3e9fb11b53a-kube-api-access-nqf29" (OuterVolumeSpecName: "kube-api-access-nqf29") pod "0e65135f-0286-4460-ae69-e3e9fb11b53a" (UID: "0e65135f-0286-4460-ae69-e3e9fb11b53a"). InnerVolumeSpecName "kube-api-access-nqf29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.427325 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-config" (OuterVolumeSpecName: "config") pod "0e65135f-0286-4460-ae69-e3e9fb11b53a" (UID: "0e65135f-0286-4460-ae69-e3e9fb11b53a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.428797 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4fb767c7-9ffa-489e-a594-433e4cb2675e" (UID: "4fb767c7-9ffa-489e-a594-433e4cb2675e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.430085 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e65135f-0286-4460-ae69-e3e9fb11b53a" (UID: "0e65135f-0286-4460-ae69-e3e9fb11b53a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.435175 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-config" (OuterVolumeSpecName: "config") pod "4fb767c7-9ffa-489e-a594-433e4cb2675e" (UID: "4fb767c7-9ffa-489e-a594-433e4cb2675e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.485671 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.485702 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.485714 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e65135f-0286-4460-ae69-e3e9fb11b53a-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.485726 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqf29\" (UniqueName: \"kubernetes.io/projected/0e65135f-0286-4460-ae69-e3e9fb11b53a-kube-api-access-nqf29\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.485741 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqf24\" (UniqueName: \"kubernetes.io/projected/4fb767c7-9ffa-489e-a594-433e4cb2675e-kube-api-access-lqf24\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.485753 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fb767c7-9ffa-489e-a594-433e4cb2675e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.633059 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" event={"ID":"4fb767c7-9ffa-489e-a594-433e4cb2675e","Type":"ContainerDied","Data":"e37392125d67ffb56a82e2e8be4467040c6a54d2a1f7ce5066b1220dcf68352d"} Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.633110 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-6rw87" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.633138 4744 scope.go:117] "RemoveContainer" containerID="befbd6837b723fccd50c63367c4d50be1e3d1e7aa1b33893167f0156fa613e00" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.635722 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.636184 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kzbvl" event={"ID":"0e65135f-0286-4460-ae69-e3e9fb11b53a","Type":"ContainerDied","Data":"403c374edb413f37a397db2034e3e94ae92bc3f3dfcf6da5d541f03dc7c129c8"} Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.637684 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" event={"ID":"403a172d-1da3-49fe-b553-5f930ec31720","Type":"ContainerStarted","Data":"fa39b7a54b30530eaf7d5d6b7050738185bd04f53ed8bf2e1c71a4cd14c769be"} Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.639774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4nncr" event={"ID":"efccc7e5-6075-43fb-b36e-15c4734e8f97","Type":"ContainerStarted","Data":"7fab1bf3db64ec3d5bdf0a3f8c13ed5600e5ade65ae31c01df58aec6ab918602"} Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.684546 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6rw87"] Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.690891 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-6rw87"] Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.702061 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kzbvl"] Dec 01 08:33:53 crc kubenswrapper[4744]: I1201 08:33:53.708127 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kzbvl"] Dec 01 08:33:54 crc kubenswrapper[4744]: I1201 08:33:54.295884 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e65135f-0286-4460-ae69-e3e9fb11b53a" path="/var/lib/kubelet/pods/0e65135f-0286-4460-ae69-e3e9fb11b53a/volumes" Dec 01 08:33:54 crc kubenswrapper[4744]: I1201 08:33:54.297024 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fb767c7-9ffa-489e-a594-433e4cb2675e" path="/var/lib/kubelet/pods/4fb767c7-9ffa-489e-a594-433e4cb2675e/volumes" Dec 01 08:33:54 crc kubenswrapper[4744]: I1201 08:33:54.608808 4744 scope.go:117] "RemoveContainer" containerID="5f1b8d0d04914a5912e725e99f62a62d4dec2df8dd5aba6413956122e9c539c7" Dec 01 08:33:55 crc kubenswrapper[4744]: I1201 08:33:55.029919 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-v49zx"] Dec 01 08:33:56 crc kubenswrapper[4744]: I1201 08:33:56.686655 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-v49zx" event={"ID":"e19eb329-ec6d-480e-a9dc-92b2b8dcc131","Type":"ContainerStarted","Data":"d405435da89ca5d12ad174440f371eb0da8e67af9185fd3a376d085b28682b60"} Dec 01 08:33:56 crc kubenswrapper[4744]: I1201 08:33:56.708676 4744 scope.go:117] "RemoveContainer" containerID="2527d9b8b6db2d7133930c1c33da958e7f29241febb320daa90a99ae4cc58481" Dec 01 08:33:58 crc kubenswrapper[4744]: I1201 08:33:58.121651 4744 scope.go:117] "RemoveContainer" containerID="5fedfc92859d2953ed3168e011a2cb520427590d0ba30c081b9f8efc81822e22" Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.401346 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.724559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pgp2f" event={"ID":"ac7e9e72-1bca-42aa-a062-e9743651f2aa","Type":"ContainerStarted","Data":"ea1654f784502739c5820c357adbee9d61a1ee7ed5f0d9eeddc97892525b4bfc"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.724680 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-pgp2f" Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.726022 4744 generic.go:334] "Generic (PLEG): container finished" podID="403a172d-1da3-49fe-b553-5f930ec31720" containerID="96ac0fc40cb6d21f376ebe88dcfd067d3dacab21446f6a8d6787598a3afd11ca" exitCode=0 Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.726090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" event={"ID":"403a172d-1da3-49fe-b553-5f930ec31720","Type":"ContainerDied","Data":"96ac0fc40cb6d21f376ebe88dcfd067d3dacab21446f6a8d6787598a3afd11ca"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.727823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6ebb51f-ba73-417a-b6a8-ff31ff555108","Type":"ContainerStarted","Data":"9ff5cbd5025d674a859e1030e1cbed59ab05b377cc6180cfdddaf5865c8a4278"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.727859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6ebb51f-ba73-417a-b6a8-ff31ff555108","Type":"ContainerStarted","Data":"47c55bd8fb3ae9652b16d6084ce75564da191373e0f628ee8b550005cdb0bffd"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.729503 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"916bb441-60a0-4452-b2cf-5b98fb9aea9f","Type":"ContainerStarted","Data":"a80ad6dc02694fbf22e02071afc6772908ace17e7de9665f772227880ea97f0c"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.729533 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"916bb441-60a0-4452-b2cf-5b98fb9aea9f","Type":"ContainerStarted","Data":"98b0958847e541a02f25f9943e825ce9bba08d36e45bbc9c145879dd5339bc0f"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.732127 4744 generic.go:334] "Generic (PLEG): container finished" podID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerID="6d7a132512589b19cdbb339129b64fc6b547cdabf65f3f01841e3a7d19c8a2ea" exitCode=0 Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.732229 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-v49zx" event={"ID":"e19eb329-ec6d-480e-a9dc-92b2b8dcc131","Type":"ContainerDied","Data":"6d7a132512589b19cdbb339129b64fc6b547cdabf65f3f01841e3a7d19c8a2ea"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.741523 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c31db90c-376f-45b5-9d13-73bcf4962135","Type":"ContainerStarted","Data":"eefad11622d7ded29cb0b17c6ddf29853cac24c52d4f3c433b577861d6a308a1"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.743821 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b711f28-53bd-46e8-9a86-44be43bd78d3","Type":"ContainerStarted","Data":"f9c22c6f4b2a44087a849cf57f1cb7a93497aefc93aaa68c7810bb6158cb01b3"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.747136 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-pgp2f" podStartSLOduration=5.580542955 podStartE2EDuration="16.747119953s" podCreationTimestamp="2025-12-01 08:33:44 +0000 UTC" firstStartedPulling="2025-12-01 08:33:48.649340573 +0000 UTC m=+960.638398494" lastFinishedPulling="2025-12-01 08:33:59.815917541 +0000 UTC m=+971.804975492" observedRunningTime="2025-12-01 08:34:00.742419069 +0000 UTC m=+972.731476990" watchObservedRunningTime="2025-12-01 08:34:00.747119953 +0000 UTC m=+972.736177874" Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.766496 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.454619257 podStartE2EDuration="15.766480039s" podCreationTimestamp="2025-12-01 08:33:45 +0000 UTC" firstStartedPulling="2025-12-01 08:33:49.517776983 +0000 UTC m=+961.506834904" lastFinishedPulling="2025-12-01 08:33:59.829637755 +0000 UTC m=+971.818695686" observedRunningTime="2025-12-01 08:34:00.76057072 +0000 UTC m=+972.749628661" watchObservedRunningTime="2025-12-01 08:34:00.766480039 +0000 UTC m=+972.755537960" Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.767534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4nncr" event={"ID":"efccc7e5-6075-43fb-b36e-15c4734e8f97","Type":"ContainerStarted","Data":"a7ce26246705d1a008a9d6bbd06f75c7b4ca705e4f675523872debc4e5329b9b"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.772028 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0153c296-2e70-43fa-90ae-bb007406d2c0","Type":"ContainerStarted","Data":"c87ca5f411966e4c8b2bbe22a7d2649c3aaeac5bb2ab932f1bd779b6d43dab4e"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.772519 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.778791 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b14f48b-6456-4329-97fa-e81ec05e2644" containerID="b379079ad871669ae4d6011425f3feb4e4c10cbedb213865531464f427b5bfaf" exitCode=0 Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.778861 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fzxwj" event={"ID":"2b14f48b-6456-4329-97fa-e81ec05e2644","Type":"ContainerDied","Data":"b379079ad871669ae4d6011425f3feb4e4c10cbedb213865531464f427b5bfaf"} Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.851176 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.257857624 podStartE2EDuration="14.851161351s" podCreationTimestamp="2025-12-01 08:33:46 +0000 UTC" firstStartedPulling="2025-12-01 08:33:48.714032101 +0000 UTC m=+960.703090012" lastFinishedPulling="2025-12-01 08:33:58.307335818 +0000 UTC m=+970.296393739" observedRunningTime="2025-12-01 08:34:00.841795802 +0000 UTC m=+972.830853723" watchObservedRunningTime="2025-12-01 08:34:00.851161351 +0000 UTC m=+972.840219272" Dec 01 08:34:00 crc kubenswrapper[4744]: I1201 08:34:00.901736 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-4nncr" podStartSLOduration=3.156245569 podStartE2EDuration="9.901717883s" podCreationTimestamp="2025-12-01 08:33:51 +0000 UTC" firstStartedPulling="2025-12-01 08:33:53.184240855 +0000 UTC m=+965.173298816" lastFinishedPulling="2025-12-01 08:33:59.929713169 +0000 UTC m=+971.918771130" observedRunningTime="2025-12-01 08:34:00.898348666 +0000 UTC m=+972.887406587" watchObservedRunningTime="2025-12-01 08:34:00.901717883 +0000 UTC m=+972.890775804" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.788331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"788a0c2b-cd13-48cb-99af-ba838e154525","Type":"ContainerStarted","Data":"3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f"} Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.791384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bf94bcff-f22a-4a35-beca-0096f08ee1f6","Type":"ContainerStarted","Data":"cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d"} Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.794478 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-v49zx" event={"ID":"e19eb329-ec6d-480e-a9dc-92b2b8dcc131","Type":"ContainerStarted","Data":"b93baf2ded278cb1ba5a6756e714ffa5d0e48e1f5a0bbade056db46b165f01c0"} Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.794840 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.796127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" event={"ID":"403a172d-1da3-49fe-b553-5f930ec31720","Type":"ContainerStarted","Data":"bcec6b8c697b90fbb190b567db54a6550ca27f70578b8544de6f20e257bb1c10"} Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.796276 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.799029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fzxwj" event={"ID":"2b14f48b-6456-4329-97fa-e81ec05e2644","Type":"ContainerStarted","Data":"7a2e1cc95147cc969f92c5f0339be921d1e100ee5e641585cf1cd2e80d79b43e"} Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.799074 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.799086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fzxwj" event={"ID":"2b14f48b-6456-4329-97fa-e81ec05e2644","Type":"ContainerStarted","Data":"ab3757e804d5fd564f9b049f4d58353d35f764159e9429f94e9e81df61735195"} Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.801465 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.820234 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.989029718 podStartE2EDuration="23.820212439s" podCreationTimestamp="2025-12-01 08:33:38 +0000 UTC" firstStartedPulling="2025-12-01 08:33:48.301329319 +0000 UTC m=+960.290387240" lastFinishedPulling="2025-12-01 08:33:57.13251203 +0000 UTC m=+969.121569961" observedRunningTime="2025-12-01 08:34:00.97718738 +0000 UTC m=+972.966245301" watchObservedRunningTime="2025-12-01 08:34:01.820212439 +0000 UTC m=+973.809270370" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.832990 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" podStartSLOduration=10.832972696 podStartE2EDuration="10.832972696s" podCreationTimestamp="2025-12-01 08:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:34:01.830848695 +0000 UTC m=+973.819906616" watchObservedRunningTime="2025-12-01 08:34:01.832972696 +0000 UTC m=+973.822030617" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.882097 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-v49zx" podStartSLOduration=10.882078646 podStartE2EDuration="10.882078646s" podCreationTimestamp="2025-12-01 08:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:34:01.876370322 +0000 UTC m=+973.865428243" watchObservedRunningTime="2025-12-01 08:34:01.882078646 +0000 UTC m=+973.871136587" Dec 01 08:34:01 crc kubenswrapper[4744]: I1201 08:34:01.895192 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-fzxwj" podStartSLOduration=9.403936043 podStartE2EDuration="17.895176552s" podCreationTimestamp="2025-12-01 08:33:44 +0000 UTC" firstStartedPulling="2025-12-01 08:33:48.641157368 +0000 UTC m=+960.630215289" lastFinishedPulling="2025-12-01 08:33:57.132397857 +0000 UTC m=+969.121455798" observedRunningTime="2025-12-01 08:34:01.891227049 +0000 UTC m=+973.880284990" watchObservedRunningTime="2025-12-01 08:34:01.895176552 +0000 UTC m=+973.884234473" Dec 01 08:34:02 crc kubenswrapper[4744]: I1201 08:34:02.219189 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 08:34:02 crc kubenswrapper[4744]: I1201 08:34:02.219513 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 08:34:02 crc kubenswrapper[4744]: I1201 08:34:02.414563 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 08:34:02 crc kubenswrapper[4744]: I1201 08:34:02.414637 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 08:34:04 crc kubenswrapper[4744]: I1201 08:34:04.837810 4744 generic.go:334] "Generic (PLEG): container finished" podID="c31db90c-376f-45b5-9d13-73bcf4962135" containerID="eefad11622d7ded29cb0b17c6ddf29853cac24c52d4f3c433b577861d6a308a1" exitCode=0 Dec 01 08:34:04 crc kubenswrapper[4744]: I1201 08:34:04.838337 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c31db90c-376f-45b5-9d13-73bcf4962135","Type":"ContainerDied","Data":"eefad11622d7ded29cb0b17c6ddf29853cac24c52d4f3c433b577861d6a308a1"} Dec 01 08:34:04 crc kubenswrapper[4744]: I1201 08:34:04.848675 4744 generic.go:334] "Generic (PLEG): container finished" podID="0b711f28-53bd-46e8-9a86-44be43bd78d3" containerID="f9c22c6f4b2a44087a849cf57f1cb7a93497aefc93aaa68c7810bb6158cb01b3" exitCode=0 Dec 01 08:34:04 crc kubenswrapper[4744]: I1201 08:34:04.848724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b711f28-53bd-46e8-9a86-44be43bd78d3","Type":"ContainerDied","Data":"f9c22c6f4b2a44087a849cf57f1cb7a93497aefc93aaa68c7810bb6158cb01b3"} Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.283470 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.352731 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.467528 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.511188 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.860674 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 08:34:05 crc kubenswrapper[4744]: E1201 08:34:05.861326 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerName="dnsmasq-dns" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.861342 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerName="dnsmasq-dns" Dec 01 08:34:05 crc kubenswrapper[4744]: E1201 08:34:05.861367 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerName="init" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.861375 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerName="init" Dec 01 08:34:05 crc kubenswrapper[4744]: E1201 08:34:05.861390 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerName="init" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.861397 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerName="init" Dec 01 08:34:05 crc kubenswrapper[4744]: E1201 08:34:05.861438 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerName="dnsmasq-dns" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.861446 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerName="dnsmasq-dns" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.861627 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fb767c7-9ffa-489e-a594-433e4cb2675e" containerName="dnsmasq-dns" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.861642 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e65135f-0286-4460-ae69-e3e9fb11b53a" containerName="dnsmasq-dns" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.862591 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.865530 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.865707 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-w6z76" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.865530 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.867108 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 08:34:05 crc kubenswrapper[4744]: I1201 08:34:05.873421 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.065126 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qklgd\" (UniqueName: \"kubernetes.io/projected/551c38c7-428c-427a-b9b3-1322157a08df-kube-api-access-qklgd\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.065293 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/551c38c7-428c-427a-b9b3-1322157a08df-config\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.065326 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.065351 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.065457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/551c38c7-428c-427a-b9b3-1322157a08df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.065574 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.065650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/551c38c7-428c-427a-b9b3-1322157a08df-scripts\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.167637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.167821 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/551c38c7-428c-427a-b9b3-1322157a08df-scripts\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.168435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qklgd\" (UniqueName: \"kubernetes.io/projected/551c38c7-428c-427a-b9b3-1322157a08df-kube-api-access-qklgd\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.168486 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/551c38c7-428c-427a-b9b3-1322157a08df-config\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.168534 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.168558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.168645 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/551c38c7-428c-427a-b9b3-1322157a08df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.169307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/551c38c7-428c-427a-b9b3-1322157a08df-scripts\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.169325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/551c38c7-428c-427a-b9b3-1322157a08df-config\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.169569 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/551c38c7-428c-427a-b9b3-1322157a08df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.173237 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.177034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.177700 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/551c38c7-428c-427a-b9b3-1322157a08df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.183771 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qklgd\" (UniqueName: \"kubernetes.io/projected/551c38c7-428c-427a-b9b3-1322157a08df-kube-api-access-qklgd\") pod \"ovn-northd-0\" (UID: \"551c38c7-428c-427a-b9b3-1322157a08df\") " pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.191281 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.632639 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.665887 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 08:34:06 crc kubenswrapper[4744]: W1201 08:34:06.676203 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod551c38c7_428c_427a_b9b3_1322157a08df.slice/crio-31aa289f382a149a48561bfcaf851c7c7c36d354ef947bc42cef7dde5054162b WatchSource:0}: Error finding container 31aa289f382a149a48561bfcaf851c7c7c36d354ef947bc42cef7dde5054162b: Status 404 returned error can't find the container with id 31aa289f382a149a48561bfcaf851c7c7c36d354ef947bc42cef7dde5054162b Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.869784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"551c38c7-428c-427a-b9b3-1322157a08df","Type":"ContainerStarted","Data":"31aa289f382a149a48561bfcaf851c7c7c36d354ef947bc42cef7dde5054162b"} Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.911612 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.962774 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-59wd5"] Dec 01 08:34:06 crc kubenswrapper[4744]: I1201 08:34:06.966348 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" podUID="403a172d-1da3-49fe-b553-5f930ec31720" containerName="dnsmasq-dns" containerID="cri-o://bcec6b8c697b90fbb190b567db54a6550ca27f70578b8544de6f20e257bb1c10" gracePeriod=10 Dec 01 08:34:07 crc kubenswrapper[4744]: I1201 08:34:07.880711 4744 generic.go:334] "Generic (PLEG): container finished" podID="403a172d-1da3-49fe-b553-5f930ec31720" containerID="bcec6b8c697b90fbb190b567db54a6550ca27f70578b8544de6f20e257bb1c10" exitCode=0 Dec 01 08:34:07 crc kubenswrapper[4744]: I1201 08:34:07.880778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" event={"ID":"403a172d-1da3-49fe-b553-5f930ec31720","Type":"ContainerDied","Data":"bcec6b8c697b90fbb190b567db54a6550ca27f70578b8544de6f20e257bb1c10"} Dec 01 08:34:08 crc kubenswrapper[4744]: I1201 08:34:08.762805 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.212877 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fbvs9"] Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.214639 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.237856 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fbvs9"] Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.352676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.352924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-config\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.353029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.353130 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfxs5\" (UniqueName: \"kubernetes.io/projected/3154b059-5545-416d-aca9-426f493b85f2-kube-api-access-gfxs5\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.353173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.454321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.454459 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfxs5\" (UniqueName: \"kubernetes.io/projected/3154b059-5545-416d-aca9-426f493b85f2-kube-api-access-gfxs5\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.454496 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.454554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.454589 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-config\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.455220 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.455322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.455509 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.455511 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-config\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.481121 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfxs5\" (UniqueName: \"kubernetes.io/projected/3154b059-5545-416d-aca9-426f493b85f2-kube-api-access-gfxs5\") pod \"dnsmasq-dns-b8fbc5445-fbvs9\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.533669 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.806246 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.907846 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcjqw\" (UniqueName: \"kubernetes.io/projected/403a172d-1da3-49fe-b553-5f930ec31720-kube-api-access-vcjqw\") pod \"403a172d-1da3-49fe-b553-5f930ec31720\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.907919 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-ovsdbserver-sb\") pod \"403a172d-1da3-49fe-b553-5f930ec31720\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.907938 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-dns-svc\") pod \"403a172d-1da3-49fe-b553-5f930ec31720\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.907987 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-config\") pod \"403a172d-1da3-49fe-b553-5f930ec31720\" (UID: \"403a172d-1da3-49fe-b553-5f930ec31720\") " Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.912822 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/403a172d-1da3-49fe-b553-5f930ec31720-kube-api-access-vcjqw" (OuterVolumeSpecName: "kube-api-access-vcjqw") pod "403a172d-1da3-49fe-b553-5f930ec31720" (UID: "403a172d-1da3-49fe-b553-5f930ec31720"). InnerVolumeSpecName "kube-api-access-vcjqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.915779 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c31db90c-376f-45b5-9d13-73bcf4962135","Type":"ContainerStarted","Data":"60d53270a300bee3a4edd78945cf325d2e0b149a668fbbf9acfa94965e12031e"} Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.929934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b711f28-53bd-46e8-9a86-44be43bd78d3","Type":"ContainerStarted","Data":"9be594ae16e2de44dd7bbcbc0c84effe8cc5d21916d9601db64561468815d345"} Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.944362 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" event={"ID":"403a172d-1da3-49fe-b553-5f930ec31720","Type":"ContainerDied","Data":"fa39b7a54b30530eaf7d5d6b7050738185bd04f53ed8bf2e1c71a4cd14c769be"} Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.944445 4744 scope.go:117] "RemoveContainer" containerID="bcec6b8c697b90fbb190b567db54a6550ca27f70578b8544de6f20e257bb1c10" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.944640 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.968732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "403a172d-1da3-49fe-b553-5f930ec31720" (UID: "403a172d-1da3-49fe-b553-5f930ec31720"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.972857 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "403a172d-1da3-49fe-b553-5f930ec31720" (UID: "403a172d-1da3-49fe-b553-5f930ec31720"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.981703 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-config" (OuterVolumeSpecName: "config") pod "403a172d-1da3-49fe-b553-5f930ec31720" (UID: "403a172d-1da3-49fe-b553-5f930ec31720"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:11 crc kubenswrapper[4744]: I1201 08:34:11.994887 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.53502415 podStartE2EDuration="35.994866032s" podCreationTimestamp="2025-12-01 08:33:36 +0000 UTC" firstStartedPulling="2025-12-01 08:33:47.833844424 +0000 UTC m=+959.822902345" lastFinishedPulling="2025-12-01 08:33:58.293686306 +0000 UTC m=+970.282744227" observedRunningTime="2025-12-01 08:34:11.977889524 +0000 UTC m=+983.966947445" watchObservedRunningTime="2025-12-01 08:34:11.994866032 +0000 UTC m=+983.983923953" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:11.999576 4744 scope.go:117] "RemoveContainer" containerID="96ac0fc40cb6d21f376ebe88dcfd067d3dacab21446f6a8d6787598a3afd11ca" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.011274 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcjqw\" (UniqueName: \"kubernetes.io/projected/403a172d-1da3-49fe-b553-5f930ec31720-kube-api-access-vcjqw\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.011309 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.011322 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.011333 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/403a172d-1da3-49fe-b553-5f930ec31720-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.011984 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=26.504032499 podStartE2EDuration="37.011965473s" podCreationTimestamp="2025-12-01 08:33:35 +0000 UTC" firstStartedPulling="2025-12-01 08:33:47.785752082 +0000 UTC m=+959.774810013" lastFinishedPulling="2025-12-01 08:33:58.293685046 +0000 UTC m=+970.282742987" observedRunningTime="2025-12-01 08:34:12.010059858 +0000 UTC m=+983.999117789" watchObservedRunningTime="2025-12-01 08:34:12.011965473 +0000 UTC m=+984.001023394" Dec 01 08:34:12 crc kubenswrapper[4744]: W1201 08:34:12.102215 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3154b059_5545_416d_aca9_426f493b85f2.slice/crio-454152fce33070d2d87453d309b93724f9dde744b8008de3f23989a7160dd33e WatchSource:0}: Error finding container 454152fce33070d2d87453d309b93724f9dde744b8008de3f23989a7160dd33e: Status 404 returned error can't find the container with id 454152fce33070d2d87453d309b93724f9dde744b8008de3f23989a7160dd33e Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.112482 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fbvs9"] Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.283266 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-59wd5"] Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.299506 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-59wd5"] Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.302695 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 01 08:34:12 crc kubenswrapper[4744]: E1201 08:34:12.303025 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="403a172d-1da3-49fe-b553-5f930ec31720" containerName="dnsmasq-dns" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.303041 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="403a172d-1da3-49fe-b553-5f930ec31720" containerName="dnsmasq-dns" Dec 01 08:34:12 crc kubenswrapper[4744]: E1201 08:34:12.303062 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="403a172d-1da3-49fe-b553-5f930ec31720" containerName="init" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.303068 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="403a172d-1da3-49fe-b553-5f930ec31720" containerName="init" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.303227 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="403a172d-1da3-49fe-b553-5f930ec31720" containerName="dnsmasq-dns" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.308882 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.312117 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.312485 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.312557 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-w747k" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.312552 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.333652 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.518165 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x2gd\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-kube-api-access-2x2gd\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.518449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/424071bb-0b84-421b-9858-016c60e89b30-lock\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.518470 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.518500 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.518531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/424071bb-0b84-421b-9858-016c60e89b30-cache\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.619420 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/424071bb-0b84-421b-9858-016c60e89b30-lock\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.619464 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.619496 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.619528 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/424071bb-0b84-421b-9858-016c60e89b30-cache\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.619567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x2gd\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-kube-api-access-2x2gd\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: E1201 08:34:12.619958 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:34:12 crc kubenswrapper[4744]: E1201 08:34:12.619986 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:34:12 crc kubenswrapper[4744]: E1201 08:34:12.620066 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift podName:424071bb-0b84-421b-9858-016c60e89b30 nodeName:}" failed. No retries permitted until 2025-12-01 08:34:13.120045206 +0000 UTC m=+985.109103117 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift") pod "swift-storage-0" (UID: "424071bb-0b84-421b-9858-016c60e89b30") : configmap "swift-ring-files" not found Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.620060 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.620102 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/424071bb-0b84-421b-9858-016c60e89b30-cache\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.620121 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/424071bb-0b84-421b-9858-016c60e89b30-lock\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.641903 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-l9b2z"] Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.643212 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.646509 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.648670 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.652882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x2gd\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-kube-api-access-2x2gd\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.668110 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.671399 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.693135 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-l9b2z"] Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.822554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-dispersionconf\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.822701 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-combined-ca-bundle\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.822788 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-ring-data-devices\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.822843 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbgtt\" (UniqueName: \"kubernetes.io/projected/4f171eaa-706a-47ae-be94-1d1cbd636df1-kube-api-access-rbgtt\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.822967 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-swiftconf\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.823015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-scripts\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.823079 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4f171eaa-706a-47ae-be94-1d1cbd636df1-etc-swift\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.925174 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4f171eaa-706a-47ae-be94-1d1cbd636df1-etc-swift\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.925249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-dispersionconf\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.925295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-combined-ca-bundle\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.925556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-ring-data-devices\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.925587 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbgtt\" (UniqueName: \"kubernetes.io/projected/4f171eaa-706a-47ae-be94-1d1cbd636df1-kube-api-access-rbgtt\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.925668 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-swiftconf\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.925694 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-scripts\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.926642 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-scripts\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.926973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4f171eaa-706a-47ae-be94-1d1cbd636df1-etc-swift\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.928062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-ring-data-devices\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.934276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-swiftconf\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.934507 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-dispersionconf\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.939857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-combined-ca-bundle\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.949603 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbgtt\" (UniqueName: \"kubernetes.io/projected/4f171eaa-706a-47ae-be94-1d1cbd636df1-kube-api-access-rbgtt\") pod \"swift-ring-rebalance-l9b2z\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.958915 4744 generic.go:334] "Generic (PLEG): container finished" podID="3154b059-5545-416d-aca9-426f493b85f2" containerID="332907dabf85e7f3aafa53b6e966f55977756863ce4fc655555d149f5f4dfd8e" exitCode=0 Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.959253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" event={"ID":"3154b059-5545-416d-aca9-426f493b85f2","Type":"ContainerDied","Data":"332907dabf85e7f3aafa53b6e966f55977756863ce4fc655555d149f5f4dfd8e"} Dec 01 08:34:12 crc kubenswrapper[4744]: I1201 08:34:12.959279 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" event={"ID":"3154b059-5545-416d-aca9-426f493b85f2","Type":"ContainerStarted","Data":"454152fce33070d2d87453d309b93724f9dde744b8008de3f23989a7160dd33e"} Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.012381 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.134066 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:13 crc kubenswrapper[4744]: E1201 08:34:13.135424 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:34:13 crc kubenswrapper[4744]: E1201 08:34:13.135444 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:34:13 crc kubenswrapper[4744]: E1201 08:34:13.136533 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift podName:424071bb-0b84-421b-9858-016c60e89b30 nodeName:}" failed. No retries permitted until 2025-12-01 08:34:14.136507047 +0000 UTC m=+986.125565158 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift") pod "swift-storage-0" (UID: "424071bb-0b84-421b-9858-016c60e89b30") : configmap "swift-ring-files" not found Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.522262 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-l9b2z"] Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.975137 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" event={"ID":"3154b059-5545-416d-aca9-426f493b85f2","Type":"ContainerStarted","Data":"6e14215e6365c84aca27e254ef9aca2511a6650ff02052805683526f016d096c"} Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.975714 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.977490 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"551c38c7-428c-427a-b9b3-1322157a08df","Type":"ContainerStarted","Data":"ea0cca56e483bd75cf611bee9415897b91e36efcf3adadfb67ccfb0e9565f330"} Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.977527 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"551c38c7-428c-427a-b9b3-1322157a08df","Type":"ContainerStarted","Data":"9f7b28940a4d44c3c26713c8b668069c8555ba78e99510370ad4c221057a2fac"} Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.977644 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.978733 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l9b2z" event={"ID":"4f171eaa-706a-47ae-be94-1d1cbd636df1","Type":"ContainerStarted","Data":"eb632c93a90e71e547f11a10d4ff051042c562df40a3529c28fca2766699f2f3"} Dec 01 08:34:13 crc kubenswrapper[4744]: I1201 08:34:13.999421 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" podStartSLOduration=2.999390358 podStartE2EDuration="2.999390358s" podCreationTimestamp="2025-12-01 08:34:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:34:13.997523914 +0000 UTC m=+985.986581835" watchObservedRunningTime="2025-12-01 08:34:13.999390358 +0000 UTC m=+985.988448279" Dec 01 08:34:14 crc kubenswrapper[4744]: I1201 08:34:14.048563 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.919855607 podStartE2EDuration="9.048543799s" podCreationTimestamp="2025-12-01 08:34:05 +0000 UTC" firstStartedPulling="2025-12-01 08:34:06.679707793 +0000 UTC m=+978.668765714" lastFinishedPulling="2025-12-01 08:34:12.808395975 +0000 UTC m=+984.797453906" observedRunningTime="2025-12-01 08:34:14.040867009 +0000 UTC m=+986.029924940" watchObservedRunningTime="2025-12-01 08:34:14.048543799 +0000 UTC m=+986.037601720" Dec 01 08:34:14 crc kubenswrapper[4744]: I1201 08:34:14.180778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:14 crc kubenswrapper[4744]: E1201 08:34:14.180990 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:34:14 crc kubenswrapper[4744]: E1201 08:34:14.181030 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:34:14 crc kubenswrapper[4744]: E1201 08:34:14.181093 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift podName:424071bb-0b84-421b-9858-016c60e89b30 nodeName:}" failed. No retries permitted until 2025-12-01 08:34:16.181076585 +0000 UTC m=+988.170134506 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift") pod "swift-storage-0" (UID: "424071bb-0b84-421b-9858-016c60e89b30") : configmap "swift-ring-files" not found Dec 01 08:34:14 crc kubenswrapper[4744]: I1201 08:34:14.293878 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="403a172d-1da3-49fe-b553-5f930ec31720" path="/var/lib/kubelet/pods/403a172d-1da3-49fe-b553-5f930ec31720/volumes" Dec 01 08:34:16 crc kubenswrapper[4744]: I1201 08:34:16.212905 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:16 crc kubenswrapper[4744]: E1201 08:34:16.213178 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:34:16 crc kubenswrapper[4744]: E1201 08:34:16.213217 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:34:16 crc kubenswrapper[4744]: E1201 08:34:16.213297 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift podName:424071bb-0b84-421b-9858-016c60e89b30 nodeName:}" failed. No retries permitted until 2025-12-01 08:34:20.213272095 +0000 UTC m=+992.202330056 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift") pod "swift-storage-0" (UID: "424071bb-0b84-421b-9858-016c60e89b30") : configmap "swift-ring-files" not found Dec 01 08:34:16 crc kubenswrapper[4744]: I1201 08:34:16.630314 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bc7876d45-59wd5" podUID="403a172d-1da3-49fe-b553-5f930ec31720" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: i/o timeout" Dec 01 08:34:16 crc kubenswrapper[4744]: I1201 08:34:16.774919 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 08:34:16 crc kubenswrapper[4744]: I1201 08:34:16.774973 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 08:34:18 crc kubenswrapper[4744]: I1201 08:34:18.171619 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 08:34:18 crc kubenswrapper[4744]: I1201 08:34:18.171950 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 08:34:18 crc kubenswrapper[4744]: I1201 08:34:18.250974 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 08:34:18 crc kubenswrapper[4744]: I1201 08:34:18.634394 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:34:18 crc kubenswrapper[4744]: I1201 08:34:18.634475 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:34:18 crc kubenswrapper[4744]: I1201 08:34:18.985718 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 08:34:19 crc kubenswrapper[4744]: I1201 08:34:19.092922 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 08:34:19 crc kubenswrapper[4744]: I1201 08:34:19.122831 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 08:34:20 crc kubenswrapper[4744]: I1201 08:34:20.032958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l9b2z" event={"ID":"4f171eaa-706a-47ae-be94-1d1cbd636df1","Type":"ContainerStarted","Data":"e3819dea04b5f891b8113616c51e2da24d7f245b2b2e29a7191bcd16e9dee981"} Dec 01 08:34:20 crc kubenswrapper[4744]: I1201 08:34:20.064494 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-l9b2z" podStartSLOduration=2.533754774 podStartE2EDuration="8.064479423s" podCreationTimestamp="2025-12-01 08:34:12 +0000 UTC" firstStartedPulling="2025-12-01 08:34:13.555035157 +0000 UTC m=+985.544093078" lastFinishedPulling="2025-12-01 08:34:19.085759806 +0000 UTC m=+991.074817727" observedRunningTime="2025-12-01 08:34:20.055438503 +0000 UTC m=+992.044496424" watchObservedRunningTime="2025-12-01 08:34:20.064479423 +0000 UTC m=+992.053537344" Dec 01 08:34:20 crc kubenswrapper[4744]: I1201 08:34:20.296227 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:20 crc kubenswrapper[4744]: E1201 08:34:20.296940 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:34:20 crc kubenswrapper[4744]: E1201 08:34:20.297004 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:34:20 crc kubenswrapper[4744]: E1201 08:34:20.297092 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift podName:424071bb-0b84-421b-9858-016c60e89b30 nodeName:}" failed. No retries permitted until 2025-12-01 08:34:28.297076582 +0000 UTC m=+1000.286134493 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift") pod "swift-storage-0" (UID: "424071bb-0b84-421b-9858-016c60e89b30") : configmap "swift-ring-files" not found Dec 01 08:34:21 crc kubenswrapper[4744]: I1201 08:34:21.535786 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:21 crc kubenswrapper[4744]: I1201 08:34:21.630138 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-v49zx"] Dec 01 08:34:21 crc kubenswrapper[4744]: I1201 08:34:21.630404 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-v49zx" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerName="dnsmasq-dns" containerID="cri-o://b93baf2ded278cb1ba5a6756e714ffa5d0e48e1f5a0bbade056db46b165f01c0" gracePeriod=10 Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.053881 4744 generic.go:334] "Generic (PLEG): container finished" podID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerID="b93baf2ded278cb1ba5a6756e714ffa5d0e48e1f5a0bbade056db46b165f01c0" exitCode=0 Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.054247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-v49zx" event={"ID":"e19eb329-ec6d-480e-a9dc-92b2b8dcc131","Type":"ContainerDied","Data":"b93baf2ded278cb1ba5a6756e714ffa5d0e48e1f5a0bbade056db46b165f01c0"} Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.054298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-v49zx" event={"ID":"e19eb329-ec6d-480e-a9dc-92b2b8dcc131","Type":"ContainerDied","Data":"d405435da89ca5d12ad174440f371eb0da8e67af9185fd3a376d085b28682b60"} Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.054314 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d405435da89ca5d12ad174440f371eb0da8e67af9185fd3a376d085b28682b60" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.087769 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.229863 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-sb\") pod \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.229944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-nb\") pod \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.229964 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clzxr\" (UniqueName: \"kubernetes.io/projected/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-kube-api-access-clzxr\") pod \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.230031 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-dns-svc\") pod \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.230094 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-config\") pod \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\" (UID: \"e19eb329-ec6d-480e-a9dc-92b2b8dcc131\") " Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.236011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-kube-api-access-clzxr" (OuterVolumeSpecName: "kube-api-access-clzxr") pod "e19eb329-ec6d-480e-a9dc-92b2b8dcc131" (UID: "e19eb329-ec6d-480e-a9dc-92b2b8dcc131"). InnerVolumeSpecName "kube-api-access-clzxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.285256 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e19eb329-ec6d-480e-a9dc-92b2b8dcc131" (UID: "e19eb329-ec6d-480e-a9dc-92b2b8dcc131"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.293062 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e19eb329-ec6d-480e-a9dc-92b2b8dcc131" (UID: "e19eb329-ec6d-480e-a9dc-92b2b8dcc131"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.299055 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-config" (OuterVolumeSpecName: "config") pod "e19eb329-ec6d-480e-a9dc-92b2b8dcc131" (UID: "e19eb329-ec6d-480e-a9dc-92b2b8dcc131"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.300232 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e19eb329-ec6d-480e-a9dc-92b2b8dcc131" (UID: "e19eb329-ec6d-480e-a9dc-92b2b8dcc131"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.331679 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.331715 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clzxr\" (UniqueName: \"kubernetes.io/projected/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-kube-api-access-clzxr\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.331724 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.331732 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:22 crc kubenswrapper[4744]: I1201 08:34:22.331740 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e19eb329-ec6d-480e-a9dc-92b2b8dcc131-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:23 crc kubenswrapper[4744]: I1201 08:34:23.062887 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-v49zx" Dec 01 08:34:23 crc kubenswrapper[4744]: I1201 08:34:23.105923 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-v49zx"] Dec 01 08:34:23 crc kubenswrapper[4744]: I1201 08:34:23.116049 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-v49zx"] Dec 01 08:34:24 crc kubenswrapper[4744]: I1201 08:34:24.303944 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" path="/var/lib/kubelet/pods/e19eb329-ec6d-480e-a9dc-92b2b8dcc131/volumes" Dec 01 08:34:26 crc kubenswrapper[4744]: I1201 08:34:26.095475 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f171eaa-706a-47ae-be94-1d1cbd636df1" containerID="e3819dea04b5f891b8113616c51e2da24d7f245b2b2e29a7191bcd16e9dee981" exitCode=0 Dec 01 08:34:26 crc kubenswrapper[4744]: I1201 08:34:26.095522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l9b2z" event={"ID":"4f171eaa-706a-47ae-be94-1d1cbd636df1","Type":"ContainerDied","Data":"e3819dea04b5f891b8113616c51e2da24d7f245b2b2e29a7191bcd16e9dee981"} Dec 01 08:34:26 crc kubenswrapper[4744]: I1201 08:34:26.271328 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 08:34:26 crc kubenswrapper[4744]: I1201 08:34:26.908516 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-v49zx" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: i/o timeout" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.108576 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.110598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l9b2z" event={"ID":"4f171eaa-706a-47ae-be94-1d1cbd636df1","Type":"ContainerDied","Data":"eb632c93a90e71e547f11a10d4ff051042c562df40a3529c28fca2766699f2f3"} Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.110636 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb632c93a90e71e547f11a10d4ff051042c562df40a3529c28fca2766699f2f3" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.110674 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l9b2z" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.260903 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-combined-ca-bundle\") pod \"4f171eaa-706a-47ae-be94-1d1cbd636df1\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.261220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-scripts\") pod \"4f171eaa-706a-47ae-be94-1d1cbd636df1\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.261444 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-dispersionconf\") pod \"4f171eaa-706a-47ae-be94-1d1cbd636df1\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.261592 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-swiftconf\") pod \"4f171eaa-706a-47ae-be94-1d1cbd636df1\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.261805 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4f171eaa-706a-47ae-be94-1d1cbd636df1-etc-swift\") pod \"4f171eaa-706a-47ae-be94-1d1cbd636df1\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.261983 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbgtt\" (UniqueName: \"kubernetes.io/projected/4f171eaa-706a-47ae-be94-1d1cbd636df1-kube-api-access-rbgtt\") pod \"4f171eaa-706a-47ae-be94-1d1cbd636df1\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.262146 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-ring-data-devices\") pod \"4f171eaa-706a-47ae-be94-1d1cbd636df1\" (UID: \"4f171eaa-706a-47ae-be94-1d1cbd636df1\") " Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.262943 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4f171eaa-706a-47ae-be94-1d1cbd636df1" (UID: "4f171eaa-706a-47ae-be94-1d1cbd636df1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.263026 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f171eaa-706a-47ae-be94-1d1cbd636df1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4f171eaa-706a-47ae-be94-1d1cbd636df1" (UID: "4f171eaa-706a-47ae-be94-1d1cbd636df1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.283139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f171eaa-706a-47ae-be94-1d1cbd636df1-kube-api-access-rbgtt" (OuterVolumeSpecName: "kube-api-access-rbgtt") pod "4f171eaa-706a-47ae-be94-1d1cbd636df1" (UID: "4f171eaa-706a-47ae-be94-1d1cbd636df1"). InnerVolumeSpecName "kube-api-access-rbgtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.288126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4f171eaa-706a-47ae-be94-1d1cbd636df1" (UID: "4f171eaa-706a-47ae-be94-1d1cbd636df1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.294618 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4f171eaa-706a-47ae-be94-1d1cbd636df1" (UID: "4f171eaa-706a-47ae-be94-1d1cbd636df1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.305109 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f171eaa-706a-47ae-be94-1d1cbd636df1" (UID: "4f171eaa-706a-47ae-be94-1d1cbd636df1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.324452 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-scripts" (OuterVolumeSpecName: "scripts") pod "4f171eaa-706a-47ae-be94-1d1cbd636df1" (UID: "4f171eaa-706a-47ae-be94-1d1cbd636df1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.366098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.367503 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.368303 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.368330 4744 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.368385 4744 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4f171eaa-706a-47ae-be94-1d1cbd636df1-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.368398 4744 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4f171eaa-706a-47ae-be94-1d1cbd636df1-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.368425 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbgtt\" (UniqueName: \"kubernetes.io/projected/4f171eaa-706a-47ae-be94-1d1cbd636df1-kube-api-access-rbgtt\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.368438 4744 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4f171eaa-706a-47ae-be94-1d1cbd636df1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.375921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/424071bb-0b84-421b-9858-016c60e89b30-etc-swift\") pod \"swift-storage-0\" (UID: \"424071bb-0b84-421b-9858-016c60e89b30\") " pod="openstack/swift-storage-0" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.383438 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4d38-account-create-update-d2vgj"] Dec 01 08:34:28 crc kubenswrapper[4744]: E1201 08:34:28.383861 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f171eaa-706a-47ae-be94-1d1cbd636df1" containerName="swift-ring-rebalance" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.383878 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f171eaa-706a-47ae-be94-1d1cbd636df1" containerName="swift-ring-rebalance" Dec 01 08:34:28 crc kubenswrapper[4744]: E1201 08:34:28.383898 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerName="dnsmasq-dns" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.383904 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerName="dnsmasq-dns" Dec 01 08:34:28 crc kubenswrapper[4744]: E1201 08:34:28.383922 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerName="init" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.383927 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerName="init" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.384112 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f171eaa-706a-47ae-be94-1d1cbd636df1" containerName="swift-ring-rebalance" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.384125 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e19eb329-ec6d-480e-a9dc-92b2b8dcc131" containerName="dnsmasq-dns" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.384686 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.387715 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.405273 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4d38-account-create-update-d2vgj"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.436772 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2c68m"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.439007 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.442492 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2c68m"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.469117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164a9dc7-b73e-48e7-973a-a5adf61e37be-operator-scripts\") pod \"keystone-db-create-2c68m\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.469156 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f627e1b2-a557-4531-bec4-a742a3f236d3-operator-scripts\") pod \"keystone-4d38-account-create-update-d2vgj\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.469183 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkd4p\" (UniqueName: \"kubernetes.io/projected/f627e1b2-a557-4531-bec4-a742a3f236d3-kube-api-access-dkd4p\") pod \"keystone-4d38-account-create-update-d2vgj\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.469222 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldxll\" (UniqueName: \"kubernetes.io/projected/164a9dc7-b73e-48e7-973a-a5adf61e37be-kube-api-access-ldxll\") pod \"keystone-db-create-2c68m\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.569941 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164a9dc7-b73e-48e7-973a-a5adf61e37be-operator-scripts\") pod \"keystone-db-create-2c68m\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.570260 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f627e1b2-a557-4531-bec4-a742a3f236d3-operator-scripts\") pod \"keystone-4d38-account-create-update-d2vgj\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.570383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkd4p\" (UniqueName: \"kubernetes.io/projected/f627e1b2-a557-4531-bec4-a742a3f236d3-kube-api-access-dkd4p\") pod \"keystone-4d38-account-create-update-d2vgj\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.570507 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldxll\" (UniqueName: \"kubernetes.io/projected/164a9dc7-b73e-48e7-973a-a5adf61e37be-kube-api-access-ldxll\") pod \"keystone-db-create-2c68m\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.571623 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f627e1b2-a557-4531-bec4-a742a3f236d3-operator-scripts\") pod \"keystone-4d38-account-create-update-d2vgj\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.571909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164a9dc7-b73e-48e7-973a-a5adf61e37be-operator-scripts\") pod \"keystone-db-create-2c68m\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.585663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkd4p\" (UniqueName: \"kubernetes.io/projected/f627e1b2-a557-4531-bec4-a742a3f236d3-kube-api-access-dkd4p\") pod \"keystone-4d38-account-create-update-d2vgj\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.586711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldxll\" (UniqueName: \"kubernetes.io/projected/164a9dc7-b73e-48e7-973a-a5adf61e37be-kube-api-access-ldxll\") pod \"keystone-db-create-2c68m\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.609363 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.625562 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-b5w9k"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.626538 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.644754 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b5w9k"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.671679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkntj\" (UniqueName: \"kubernetes.io/projected/6ba566b4-fecf-4538-b02d-32b1173b0cd2-kube-api-access-hkntj\") pod \"placement-db-create-b5w9k\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.672039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ba566b4-fecf-4538-b02d-32b1173b0cd2-operator-scripts\") pod \"placement-db-create-b5w9k\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.739719 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.747923 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8b45-account-create-update-tqk55"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.749117 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.752854 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.758448 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.759305 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8b45-account-create-update-tqk55"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.773016 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pplkn\" (UniqueName: \"kubernetes.io/projected/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-kube-api-access-pplkn\") pod \"placement-8b45-account-create-update-tqk55\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.773081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ba566b4-fecf-4538-b02d-32b1173b0cd2-operator-scripts\") pod \"placement-db-create-b5w9k\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.773208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-operator-scripts\") pod \"placement-8b45-account-create-update-tqk55\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.773263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkntj\" (UniqueName: \"kubernetes.io/projected/6ba566b4-fecf-4538-b02d-32b1173b0cd2-kube-api-access-hkntj\") pod \"placement-db-create-b5w9k\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.773836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ba566b4-fecf-4538-b02d-32b1173b0cd2-operator-scripts\") pod \"placement-db-create-b5w9k\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.797721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkntj\" (UniqueName: \"kubernetes.io/projected/6ba566b4-fecf-4538-b02d-32b1173b0cd2-kube-api-access-hkntj\") pod \"placement-db-create-b5w9k\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.874325 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-operator-scripts\") pod \"placement-8b45-account-create-update-tqk55\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.874630 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pplkn\" (UniqueName: \"kubernetes.io/projected/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-kube-api-access-pplkn\") pod \"placement-8b45-account-create-update-tqk55\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.875606 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-operator-scripts\") pod \"placement-8b45-account-create-update-tqk55\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.894290 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pplkn\" (UniqueName: \"kubernetes.io/projected/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-kube-api-access-pplkn\") pod \"placement-8b45-account-create-update-tqk55\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.927558 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rz4d8"] Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.928831 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.940447 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:28 crc kubenswrapper[4744]: I1201 08:34:28.943383 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rz4d8"] Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.055146 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-3282-account-create-update-q8hrf"] Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.056436 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.058894 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.061623 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3282-account-create-update-q8hrf"] Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.078347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-879dm\" (UniqueName: \"kubernetes.io/projected/a9476e9f-b34f-40b4-96d1-744e12762c72-kube-api-access-879dm\") pod \"glance-db-create-rz4d8\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.078551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9476e9f-b34f-40b4-96d1-744e12762c72-operator-scripts\") pod \"glance-db-create-rz4d8\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.078684 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.180112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9476e9f-b34f-40b4-96d1-744e12762c72-operator-scripts\") pod \"glance-db-create-rz4d8\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.180152 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-879dm\" (UniqueName: \"kubernetes.io/projected/a9476e9f-b34f-40b4-96d1-744e12762c72-kube-api-access-879dm\") pod \"glance-db-create-rz4d8\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.180208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls29t\" (UniqueName: \"kubernetes.io/projected/b84980f5-3ba1-4b30-b572-4401e5487373-kube-api-access-ls29t\") pod \"glance-3282-account-create-update-q8hrf\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.180302 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b84980f5-3ba1-4b30-b572-4401e5487373-operator-scripts\") pod \"glance-3282-account-create-update-q8hrf\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.181188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9476e9f-b34f-40b4-96d1-744e12762c72-operator-scripts\") pod \"glance-db-create-rz4d8\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.200010 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-879dm\" (UniqueName: \"kubernetes.io/projected/a9476e9f-b34f-40b4-96d1-744e12762c72-kube-api-access-879dm\") pod \"glance-db-create-rz4d8\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.231756 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4d38-account-create-update-d2vgj"] Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.244044 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.247648 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2c68m"] Dec 01 08:34:29 crc kubenswrapper[4744]: W1201 08:34:29.248544 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf627e1b2_a557_4531_bec4_a742a3f236d3.slice/crio-5be6cfc970e15ba7b4f52899b15ac9cda1b6e6fbb03a243a785d21d854041259 WatchSource:0}: Error finding container 5be6cfc970e15ba7b4f52899b15ac9cda1b6e6fbb03a243a785d21d854041259: Status 404 returned error can't find the container with id 5be6cfc970e15ba7b4f52899b15ac9cda1b6e6fbb03a243a785d21d854041259 Dec 01 08:34:29 crc kubenswrapper[4744]: W1201 08:34:29.250931 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod424071bb_0b84_421b_9858_016c60e89b30.slice/crio-555036575efe3396cc080b880a6645246dd95f084b68faf353d16c5139d22f5d WatchSource:0}: Error finding container 555036575efe3396cc080b880a6645246dd95f084b68faf353d16c5139d22f5d: Status 404 returned error can't find the container with id 555036575efe3396cc080b880a6645246dd95f084b68faf353d16c5139d22f5d Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.252750 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.282159 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b84980f5-3ba1-4b30-b572-4401e5487373-operator-scripts\") pod \"glance-3282-account-create-update-q8hrf\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.282229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls29t\" (UniqueName: \"kubernetes.io/projected/b84980f5-3ba1-4b30-b572-4401e5487373-kube-api-access-ls29t\") pod \"glance-3282-account-create-update-q8hrf\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.283086 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b84980f5-3ba1-4b30-b572-4401e5487373-operator-scripts\") pod \"glance-3282-account-create-update-q8hrf\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.298626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls29t\" (UniqueName: \"kubernetes.io/projected/b84980f5-3ba1-4b30-b572-4401e5487373-kube-api-access-ls29t\") pod \"glance-3282-account-create-update-q8hrf\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.399789 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.403822 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b5w9k"] Dec 01 08:34:29 crc kubenswrapper[4744]: W1201 08:34:29.433704 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ba566b4_fecf_4538_b02d_32b1173b0cd2.slice/crio-dda0a500785a812500d78a8c4a5fe59c08ee53173d25d3945758a5d252df8671 WatchSource:0}: Error finding container dda0a500785a812500d78a8c4a5fe59c08ee53173d25d3945758a5d252df8671: Status 404 returned error can't find the container with id dda0a500785a812500d78a8c4a5fe59c08ee53173d25d3945758a5d252df8671 Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.546172 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8b45-account-create-update-tqk55"] Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.682250 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rz4d8"] Dec 01 08:34:29 crc kubenswrapper[4744]: I1201 08:34:29.836159 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3282-account-create-update-q8hrf"] Dec 01 08:34:29 crc kubenswrapper[4744]: W1201 08:34:29.897535 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb84980f5_3ba1_4b30_b572_4401e5487373.slice/crio-ac7b5e3a255c153edde02d1ceb64cc843de47a7eb03620eaba69ffa72b53d6ab WatchSource:0}: Error finding container ac7b5e3a255c153edde02d1ceb64cc843de47a7eb03620eaba69ffa72b53d6ab: Status 404 returned error can't find the container with id ac7b5e3a255c153edde02d1ceb64cc843de47a7eb03620eaba69ffa72b53d6ab Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.126943 4744 generic.go:334] "Generic (PLEG): container finished" podID="164a9dc7-b73e-48e7-973a-a5adf61e37be" containerID="b883794a457fb19259ecefd6dac249e578dfe29a407fdb7f8f7ea0e04aba7f68" exitCode=0 Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.127124 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2c68m" event={"ID":"164a9dc7-b73e-48e7-973a-a5adf61e37be","Type":"ContainerDied","Data":"b883794a457fb19259ecefd6dac249e578dfe29a407fdb7f8f7ea0e04aba7f68"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.127228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2c68m" event={"ID":"164a9dc7-b73e-48e7-973a-a5adf61e37be","Type":"ContainerStarted","Data":"e2f5fef1b74e06faf4bb053e5bcae465a7718fd5df21618b3634dbaff72f56ec"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.129956 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3282-account-create-update-q8hrf" event={"ID":"b84980f5-3ba1-4b30-b572-4401e5487373","Type":"ContainerStarted","Data":"3cb0c1e4ef01926601f9a5747ac84e703e156cf5c434af23492a3782035580d4"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.130008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3282-account-create-update-q8hrf" event={"ID":"b84980f5-3ba1-4b30-b572-4401e5487373","Type":"ContainerStarted","Data":"ac7b5e3a255c153edde02d1ceb64cc843de47a7eb03620eaba69ffa72b53d6ab"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.133128 4744 generic.go:334] "Generic (PLEG): container finished" podID="c3a278c5-97a6-4d6a-9a08-f770d6962f5d" containerID="244c4e59c0826358c4d1b6f4173ab80c21a61adfe61dcbdc9af0a3c6bc2c265f" exitCode=0 Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.133236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8b45-account-create-update-tqk55" event={"ID":"c3a278c5-97a6-4d6a-9a08-f770d6962f5d","Type":"ContainerDied","Data":"244c4e59c0826358c4d1b6f4173ab80c21a61adfe61dcbdc9af0a3c6bc2c265f"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.133254 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8b45-account-create-update-tqk55" event={"ID":"c3a278c5-97a6-4d6a-9a08-f770d6962f5d","Type":"ContainerStarted","Data":"0f0f1be5aacdb078ea7d406d7c90cc1a85bc565378e22d04358bcef94be93a74"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.135093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rz4d8" event={"ID":"a9476e9f-b34f-40b4-96d1-744e12762c72","Type":"ContainerStarted","Data":"d3c1bff80bbaab51f7b3bb8f48ee7823284f295d0956c210eac4f8f02439dd15"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.135199 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rz4d8" event={"ID":"a9476e9f-b34f-40b4-96d1-744e12762c72","Type":"ContainerStarted","Data":"2c7fecae99e3607db608a621d0372cdacf524a707806c7b376291d1730d140a0"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.136805 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"555036575efe3396cc080b880a6645246dd95f084b68faf353d16c5139d22f5d"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.139842 4744 generic.go:334] "Generic (PLEG): container finished" podID="f627e1b2-a557-4531-bec4-a742a3f236d3" containerID="9ba871b26aa46f5224369b95b55187b2a66ebac58b9ec25ac192bd761432b827" exitCode=0 Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.140021 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4d38-account-create-update-d2vgj" event={"ID":"f627e1b2-a557-4531-bec4-a742a3f236d3","Type":"ContainerDied","Data":"9ba871b26aa46f5224369b95b55187b2a66ebac58b9ec25ac192bd761432b827"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.140071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4d38-account-create-update-d2vgj" event={"ID":"f627e1b2-a557-4531-bec4-a742a3f236d3","Type":"ContainerStarted","Data":"5be6cfc970e15ba7b4f52899b15ac9cda1b6e6fbb03a243a785d21d854041259"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.146335 4744 generic.go:334] "Generic (PLEG): container finished" podID="6ba566b4-fecf-4538-b02d-32b1173b0cd2" containerID="a07782f241b0eb668936c74ff218d2d4b1d078ea4cf4a82909330cc1fcb52c2d" exitCode=0 Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.146456 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b5w9k" event={"ID":"6ba566b4-fecf-4538-b02d-32b1173b0cd2","Type":"ContainerDied","Data":"a07782f241b0eb668936c74ff218d2d4b1d078ea4cf4a82909330cc1fcb52c2d"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.146494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b5w9k" event={"ID":"6ba566b4-fecf-4538-b02d-32b1173b0cd2","Type":"ContainerStarted","Data":"dda0a500785a812500d78a8c4a5fe59c08ee53173d25d3945758a5d252df8671"} Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.186037 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-3282-account-create-update-q8hrf" podStartSLOduration=1.186017979 podStartE2EDuration="1.186017979s" podCreationTimestamp="2025-12-01 08:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:34:30.17908413 +0000 UTC m=+1002.168142071" watchObservedRunningTime="2025-12-01 08:34:30.186017979 +0000 UTC m=+1002.175075910" Dec 01 08:34:30 crc kubenswrapper[4744]: I1201 08:34:30.225872 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-rz4d8" podStartSLOduration=2.225847783 podStartE2EDuration="2.225847783s" podCreationTimestamp="2025-12-01 08:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:34:30.213338114 +0000 UTC m=+1002.202396045" watchObservedRunningTime="2025-12-01 08:34:30.225847783 +0000 UTC m=+1002.214905724" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.167563 4744 generic.go:334] "Generic (PLEG): container finished" podID="a9476e9f-b34f-40b4-96d1-744e12762c72" containerID="d3c1bff80bbaab51f7b3bb8f48ee7823284f295d0956c210eac4f8f02439dd15" exitCode=0 Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.167907 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rz4d8" event={"ID":"a9476e9f-b34f-40b4-96d1-744e12762c72","Type":"ContainerDied","Data":"d3c1bff80bbaab51f7b3bb8f48ee7823284f295d0956c210eac4f8f02439dd15"} Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.173110 4744 generic.go:334] "Generic (PLEG): container finished" podID="b84980f5-3ba1-4b30-b572-4401e5487373" containerID="3cb0c1e4ef01926601f9a5747ac84e703e156cf5c434af23492a3782035580d4" exitCode=0 Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.173242 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3282-account-create-update-q8hrf" event={"ID":"b84980f5-3ba1-4b30-b572-4401e5487373","Type":"ContainerDied","Data":"3cb0c1e4ef01926601f9a5747ac84e703e156cf5c434af23492a3782035580d4"} Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.810662 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.820702 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.877864 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.884924 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.927212 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ba566b4-fecf-4538-b02d-32b1173b0cd2-operator-scripts\") pod \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.927280 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkntj\" (UniqueName: \"kubernetes.io/projected/6ba566b4-fecf-4538-b02d-32b1173b0cd2-kube-api-access-hkntj\") pod \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\" (UID: \"6ba566b4-fecf-4538-b02d-32b1173b0cd2\") " Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.927304 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f627e1b2-a557-4531-bec4-a742a3f236d3-operator-scripts\") pod \"f627e1b2-a557-4531-bec4-a742a3f236d3\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.927333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkd4p\" (UniqueName: \"kubernetes.io/projected/f627e1b2-a557-4531-bec4-a742a3f236d3-kube-api-access-dkd4p\") pod \"f627e1b2-a557-4531-bec4-a742a3f236d3\" (UID: \"f627e1b2-a557-4531-bec4-a742a3f236d3\") " Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.928657 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ba566b4-fecf-4538-b02d-32b1173b0cd2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ba566b4-fecf-4538-b02d-32b1173b0cd2" (UID: "6ba566b4-fecf-4538-b02d-32b1173b0cd2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.929246 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f627e1b2-a557-4531-bec4-a742a3f236d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f627e1b2-a557-4531-bec4-a742a3f236d3" (UID: "f627e1b2-a557-4531-bec4-a742a3f236d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.939086 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f627e1b2-a557-4531-bec4-a742a3f236d3-kube-api-access-dkd4p" (OuterVolumeSpecName: "kube-api-access-dkd4p") pod "f627e1b2-a557-4531-bec4-a742a3f236d3" (UID: "f627e1b2-a557-4531-bec4-a742a3f236d3"). InnerVolumeSpecName "kube-api-access-dkd4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:31 crc kubenswrapper[4744]: I1201 08:34:31.939115 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba566b4-fecf-4538-b02d-32b1173b0cd2-kube-api-access-hkntj" (OuterVolumeSpecName: "kube-api-access-hkntj") pod "6ba566b4-fecf-4538-b02d-32b1173b0cd2" (UID: "6ba566b4-fecf-4538-b02d-32b1173b0cd2"). InnerVolumeSpecName "kube-api-access-hkntj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.028474 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-operator-scripts\") pod \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.028541 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldxll\" (UniqueName: \"kubernetes.io/projected/164a9dc7-b73e-48e7-973a-a5adf61e37be-kube-api-access-ldxll\") pod \"164a9dc7-b73e-48e7-973a-a5adf61e37be\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.028640 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164a9dc7-b73e-48e7-973a-a5adf61e37be-operator-scripts\") pod \"164a9dc7-b73e-48e7-973a-a5adf61e37be\" (UID: \"164a9dc7-b73e-48e7-973a-a5adf61e37be\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.028750 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pplkn\" (UniqueName: \"kubernetes.io/projected/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-kube-api-access-pplkn\") pod \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\" (UID: \"c3a278c5-97a6-4d6a-9a08-f770d6962f5d\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.029132 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ba566b4-fecf-4538-b02d-32b1173b0cd2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.029188 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkntj\" (UniqueName: \"kubernetes.io/projected/6ba566b4-fecf-4538-b02d-32b1173b0cd2-kube-api-access-hkntj\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.029198 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f627e1b2-a557-4531-bec4-a742a3f236d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.029207 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkd4p\" (UniqueName: \"kubernetes.io/projected/f627e1b2-a557-4531-bec4-a742a3f236d3-kube-api-access-dkd4p\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.030339 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c3a278c5-97a6-4d6a-9a08-f770d6962f5d" (UID: "c3a278c5-97a6-4d6a-9a08-f770d6962f5d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.030495 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/164a9dc7-b73e-48e7-973a-a5adf61e37be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "164a9dc7-b73e-48e7-973a-a5adf61e37be" (UID: "164a9dc7-b73e-48e7-973a-a5adf61e37be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.032861 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-kube-api-access-pplkn" (OuterVolumeSpecName: "kube-api-access-pplkn") pod "c3a278c5-97a6-4d6a-9a08-f770d6962f5d" (UID: "c3a278c5-97a6-4d6a-9a08-f770d6962f5d"). InnerVolumeSpecName "kube-api-access-pplkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.034505 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/164a9dc7-b73e-48e7-973a-a5adf61e37be-kube-api-access-ldxll" (OuterVolumeSpecName: "kube-api-access-ldxll") pod "164a9dc7-b73e-48e7-973a-a5adf61e37be" (UID: "164a9dc7-b73e-48e7-973a-a5adf61e37be"). InnerVolumeSpecName "kube-api-access-ldxll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.131029 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldxll\" (UniqueName: \"kubernetes.io/projected/164a9dc7-b73e-48e7-973a-a5adf61e37be-kube-api-access-ldxll\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.131059 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.131068 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/164a9dc7-b73e-48e7-973a-a5adf61e37be-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.131077 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pplkn\" (UniqueName: \"kubernetes.io/projected/c3a278c5-97a6-4d6a-9a08-f770d6962f5d-kube-api-access-pplkn\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.181845 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"eb4954258c86026d6e6513edb49dd9e967856f03eac0dfa2abf093efc5a2ec04"} Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.181901 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"41ddf6bf34e400f12b1657829bc3fd1ef6e06bfa1e3b227ee5bb55350d32b5f8"} Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.183101 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4d38-account-create-update-d2vgj" event={"ID":"f627e1b2-a557-4531-bec4-a742a3f236d3","Type":"ContainerDied","Data":"5be6cfc970e15ba7b4f52899b15ac9cda1b6e6fbb03a243a785d21d854041259"} Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.183127 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d38-account-create-update-d2vgj" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.183140 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5be6cfc970e15ba7b4f52899b15ac9cda1b6e6fbb03a243a785d21d854041259" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.185043 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2c68m" event={"ID":"164a9dc7-b73e-48e7-973a-a5adf61e37be","Type":"ContainerDied","Data":"e2f5fef1b74e06faf4bb053e5bcae465a7718fd5df21618b3634dbaff72f56ec"} Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.185064 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2f5fef1b74e06faf4bb053e5bcae465a7718fd5df21618b3634dbaff72f56ec" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.185075 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2c68m" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.186556 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b5w9k" event={"ID":"6ba566b4-fecf-4538-b02d-32b1173b0cd2","Type":"ContainerDied","Data":"dda0a500785a812500d78a8c4a5fe59c08ee53173d25d3945758a5d252df8671"} Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.186582 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dda0a500785a812500d78a8c4a5fe59c08ee53173d25d3945758a5d252df8671" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.186612 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b5w9k" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.188031 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8b45-account-create-update-tqk55" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.191139 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8b45-account-create-update-tqk55" event={"ID":"c3a278c5-97a6-4d6a-9a08-f770d6962f5d","Type":"ContainerDied","Data":"0f0f1be5aacdb078ea7d406d7c90cc1a85bc565378e22d04358bcef94be93a74"} Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.191183 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f0f1be5aacdb078ea7d406d7c90cc1a85bc565378e22d04358bcef94be93a74" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.529731 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.609527 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.641078 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls29t\" (UniqueName: \"kubernetes.io/projected/b84980f5-3ba1-4b30-b572-4401e5487373-kube-api-access-ls29t\") pod \"b84980f5-3ba1-4b30-b572-4401e5487373\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.641144 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b84980f5-3ba1-4b30-b572-4401e5487373-operator-scripts\") pod \"b84980f5-3ba1-4b30-b572-4401e5487373\" (UID: \"b84980f5-3ba1-4b30-b572-4401e5487373\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.642868 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b84980f5-3ba1-4b30-b572-4401e5487373-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b84980f5-3ba1-4b30-b572-4401e5487373" (UID: "b84980f5-3ba1-4b30-b572-4401e5487373"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.646463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b84980f5-3ba1-4b30-b572-4401e5487373-kube-api-access-ls29t" (OuterVolumeSpecName: "kube-api-access-ls29t") pod "b84980f5-3ba1-4b30-b572-4401e5487373" (UID: "b84980f5-3ba1-4b30-b572-4401e5487373"). InnerVolumeSpecName "kube-api-access-ls29t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.742339 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-879dm\" (UniqueName: \"kubernetes.io/projected/a9476e9f-b34f-40b4-96d1-744e12762c72-kube-api-access-879dm\") pod \"a9476e9f-b34f-40b4-96d1-744e12762c72\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.742398 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9476e9f-b34f-40b4-96d1-744e12762c72-operator-scripts\") pod \"a9476e9f-b34f-40b4-96d1-744e12762c72\" (UID: \"a9476e9f-b34f-40b4-96d1-744e12762c72\") " Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.743500 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9476e9f-b34f-40b4-96d1-744e12762c72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9476e9f-b34f-40b4-96d1-744e12762c72" (UID: "a9476e9f-b34f-40b4-96d1-744e12762c72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.743543 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls29t\" (UniqueName: \"kubernetes.io/projected/b84980f5-3ba1-4b30-b572-4401e5487373-kube-api-access-ls29t\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.743562 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b84980f5-3ba1-4b30-b572-4401e5487373-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.748617 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9476e9f-b34f-40b4-96d1-744e12762c72-kube-api-access-879dm" (OuterVolumeSpecName: "kube-api-access-879dm") pod "a9476e9f-b34f-40b4-96d1-744e12762c72" (UID: "a9476e9f-b34f-40b4-96d1-744e12762c72"). InnerVolumeSpecName "kube-api-access-879dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.845720 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-879dm\" (UniqueName: \"kubernetes.io/projected/a9476e9f-b34f-40b4-96d1-744e12762c72-kube-api-access-879dm\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:32 crc kubenswrapper[4744]: I1201 08:34:32.845759 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9476e9f-b34f-40b4-96d1-744e12762c72-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.200839 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rz4d8" event={"ID":"a9476e9f-b34f-40b4-96d1-744e12762c72","Type":"ContainerDied","Data":"2c7fecae99e3607db608a621d0372cdacf524a707806c7b376291d1730d140a0"} Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.200880 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c7fecae99e3607db608a621d0372cdacf524a707806c7b376291d1730d140a0" Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.201005 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rz4d8" Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.205640 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"b6b6069e156f0b413ded15c6dd3ae1235fdd3e434dc5605b491e627a476e3f9b"} Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.205716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"00270b37a937199aaa2778a54bd1fdb18006b378850628abfceac0b92ea81e3c"} Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.208236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3282-account-create-update-q8hrf" event={"ID":"b84980f5-3ba1-4b30-b572-4401e5487373","Type":"ContainerDied","Data":"ac7b5e3a255c153edde02d1ceb64cc843de47a7eb03620eaba69ffa72b53d6ab"} Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.208286 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac7b5e3a255c153edde02d1ceb64cc843de47a7eb03620eaba69ffa72b53d6ab" Dec 01 08:34:33 crc kubenswrapper[4744]: I1201 08:34:33.208289 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3282-account-create-update-q8hrf" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.214701 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jx6b8"] Dec 01 08:34:34 crc kubenswrapper[4744]: E1201 08:34:34.215400 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f627e1b2-a557-4531-bec4-a742a3f236d3" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215436 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f627e1b2-a557-4531-bec4-a742a3f236d3" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: E1201 08:34:34.215465 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba566b4-fecf-4538-b02d-32b1173b0cd2" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215474 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba566b4-fecf-4538-b02d-32b1173b0cd2" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: E1201 08:34:34.215485 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9476e9f-b34f-40b4-96d1-744e12762c72" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215493 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9476e9f-b34f-40b4-96d1-744e12762c72" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: E1201 08:34:34.215509 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="164a9dc7-b73e-48e7-973a-a5adf61e37be" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215516 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="164a9dc7-b73e-48e7-973a-a5adf61e37be" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: E1201 08:34:34.215528 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a278c5-97a6-4d6a-9a08-f770d6962f5d" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215535 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a278c5-97a6-4d6a-9a08-f770d6962f5d" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: E1201 08:34:34.215543 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84980f5-3ba1-4b30-b572-4401e5487373" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215551 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84980f5-3ba1-4b30-b572-4401e5487373" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215764 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="164a9dc7-b73e-48e7-973a-a5adf61e37be" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215793 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f627e1b2-a557-4531-bec4-a742a3f236d3" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215810 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b84980f5-3ba1-4b30-b572-4401e5487373" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215824 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba566b4-fecf-4538-b02d-32b1173b0cd2" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215834 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a278c5-97a6-4d6a-9a08-f770d6962f5d" containerName="mariadb-account-create-update" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.215843 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9476e9f-b34f-40b4-96d1-744e12762c72" containerName="mariadb-database-create" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.216420 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.221099 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gl4sq" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.221125 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"788a0c2b-cd13-48cb-99af-ba838e154525","Type":"ContainerDied","Data":"3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f"} Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.221091 4744 generic.go:334] "Generic (PLEG): container finished" podID="788a0c2b-cd13-48cb-99af-ba838e154525" containerID="3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f" exitCode=0 Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.221428 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.225089 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerID="cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d" exitCode=0 Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.225133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bf94bcff-f22a-4a35-beca-0096f08ee1f6","Type":"ContainerDied","Data":"cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d"} Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.264916 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jx6b8"] Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.372637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-combined-ca-bundle\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.372695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9kzj\" (UniqueName: \"kubernetes.io/projected/08cb1243-7645-41a5-b5d2-ea22230908c7-kube-api-access-m9kzj\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.372842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-config-data\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.372985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-db-sync-config-data\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.475017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-db-sync-config-data\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.476092 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-combined-ca-bundle\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.476189 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9kzj\" (UniqueName: \"kubernetes.io/projected/08cb1243-7645-41a5-b5d2-ea22230908c7-kube-api-access-m9kzj\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.476260 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-config-data\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.480710 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-db-sync-config-data\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.480762 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-combined-ca-bundle\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.486089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-config-data\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.503338 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9kzj\" (UniqueName: \"kubernetes.io/projected/08cb1243-7645-41a5-b5d2-ea22230908c7-kube-api-access-m9kzj\") pod \"glance-db-sync-jx6b8\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.557236 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-pgp2f" podUID="ac7e9e72-1bca-42aa-a062-e9743651f2aa" containerName="ovn-controller" probeResult="failure" output=< Dec 01 08:34:34 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 08:34:34 crc kubenswrapper[4744]: > Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.562068 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.568094 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fzxwj" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.689270 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jx6b8" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.783679 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pgp2f-config-m7blz"] Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.784594 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.787689 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.794522 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pgp2f-config-m7blz"] Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.883895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjltk\" (UniqueName: \"kubernetes.io/projected/4cb15ea4-0788-4003-bf31-9e8bc453d778-kube-api-access-bjltk\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.884214 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-additional-scripts\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.884434 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-scripts\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.884792 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.884949 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run-ovn\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.884981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-log-ovn\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-scripts\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988428 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988474 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run-ovn\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988506 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-log-ovn\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988533 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjltk\" (UniqueName: \"kubernetes.io/projected/4cb15ea4-0788-4003-bf31-9e8bc453d778-kube-api-access-bjltk\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-additional-scripts\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run-ovn\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.988750 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.989040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-log-ovn\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.989204 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-additional-scripts\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:34 crc kubenswrapper[4744]: I1201 08:34:34.990895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-scripts\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.007931 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjltk\" (UniqueName: \"kubernetes.io/projected/4cb15ea4-0788-4003-bf31-9e8bc453d778-kube-api-access-bjltk\") pod \"ovn-controller-pgp2f-config-m7blz\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.101542 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.252028 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"50e1a8005b6189c716111c1772f8addf5d315362ba791ecb7ad78e29bc870875"} Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.252066 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"0487803dd8e0f9c7ee17825b1fd64342166205de5e6d09f0f4dc207a85ac9293"} Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.255343 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"788a0c2b-cd13-48cb-99af-ba838e154525","Type":"ContainerStarted","Data":"55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3"} Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.255644 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.263564 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bf94bcff-f22a-4a35-beca-0096f08ee1f6","Type":"ContainerStarted","Data":"2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993"} Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.263951 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.284657 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=48.178226974 podStartE2EDuration="1m1.2846381s" podCreationTimestamp="2025-12-01 08:33:34 +0000 UTC" firstStartedPulling="2025-12-01 08:33:46.803318809 +0000 UTC m=+958.792376730" lastFinishedPulling="2025-12-01 08:33:59.909729895 +0000 UTC m=+971.898787856" observedRunningTime="2025-12-01 08:34:35.282760126 +0000 UTC m=+1007.271818047" watchObservedRunningTime="2025-12-01 08:34:35.2846381 +0000 UTC m=+1007.273696021" Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.353259 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.725244002 podStartE2EDuration="1m2.353241241s" podCreationTimestamp="2025-12-01 08:33:33 +0000 UTC" firstStartedPulling="2025-12-01 08:33:48.403362659 +0000 UTC m=+960.392420600" lastFinishedPulling="2025-12-01 08:34:00.031359918 +0000 UTC m=+972.020417839" observedRunningTime="2025-12-01 08:34:35.324260238 +0000 UTC m=+1007.313318159" watchObservedRunningTime="2025-12-01 08:34:35.353241241 +0000 UTC m=+1007.342299162" Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.392960 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jx6b8"] Dec 01 08:34:35 crc kubenswrapper[4744]: W1201 08:34:35.432856 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08cb1243_7645_41a5_b5d2_ea22230908c7.slice/crio-c3b8e2dae6b6ddc8b09fdc7a0a0f79261e52d199216bb3b37be4df0a211501c0 WatchSource:0}: Error finding container c3b8e2dae6b6ddc8b09fdc7a0a0f79261e52d199216bb3b37be4df0a211501c0: Status 404 returned error can't find the container with id c3b8e2dae6b6ddc8b09fdc7a0a0f79261e52d199216bb3b37be4df0a211501c0 Dec 01 08:34:35 crc kubenswrapper[4744]: I1201 08:34:35.599934 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pgp2f-config-m7blz"] Dec 01 08:34:36 crc kubenswrapper[4744]: I1201 08:34:36.271080 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jx6b8" event={"ID":"08cb1243-7645-41a5-b5d2-ea22230908c7","Type":"ContainerStarted","Data":"c3b8e2dae6b6ddc8b09fdc7a0a0f79261e52d199216bb3b37be4df0a211501c0"} Dec 01 08:34:36 crc kubenswrapper[4744]: I1201 08:34:36.273658 4744 generic.go:334] "Generic (PLEG): container finished" podID="4cb15ea4-0788-4003-bf31-9e8bc453d778" containerID="f64967bbc5b21958514db37dd22f8dbe11703b724f23fb024f4d510562ac1cb0" exitCode=0 Dec 01 08:34:36 crc kubenswrapper[4744]: I1201 08:34:36.273801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pgp2f-config-m7blz" event={"ID":"4cb15ea4-0788-4003-bf31-9e8bc453d778","Type":"ContainerDied","Data":"f64967bbc5b21958514db37dd22f8dbe11703b724f23fb024f4d510562ac1cb0"} Dec 01 08:34:36 crc kubenswrapper[4744]: I1201 08:34:36.273851 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pgp2f-config-m7blz" event={"ID":"4cb15ea4-0788-4003-bf31-9e8bc453d778","Type":"ContainerStarted","Data":"73b8cd8dba92f48e9f8c0e14c695288a5e8441be39586ac89a7f54048057e2b8"} Dec 01 08:34:36 crc kubenswrapper[4744]: I1201 08:34:36.277496 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"20bc7aaf7d402587d229030d1178302c8903012181d25001a74e6e668a7533da"} Dec 01 08:34:36 crc kubenswrapper[4744]: I1201 08:34:36.277546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"1c31eb0b748691791034b4e42872c1172dad1e1ee219938633b7fb84f90ba243"} Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.875309 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.933884 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-additional-scripts\") pod \"4cb15ea4-0788-4003-bf31-9e8bc453d778\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.934339 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjltk\" (UniqueName: \"kubernetes.io/projected/4cb15ea4-0788-4003-bf31-9e8bc453d778-kube-api-access-bjltk\") pod \"4cb15ea4-0788-4003-bf31-9e8bc453d778\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.934401 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run\") pod \"4cb15ea4-0788-4003-bf31-9e8bc453d778\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.934449 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-log-ovn\") pod \"4cb15ea4-0788-4003-bf31-9e8bc453d778\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.934483 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-scripts\") pod \"4cb15ea4-0788-4003-bf31-9e8bc453d778\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.934520 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run-ovn\") pod \"4cb15ea4-0788-4003-bf31-9e8bc453d778\" (UID: \"4cb15ea4-0788-4003-bf31-9e8bc453d778\") " Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.935056 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4cb15ea4-0788-4003-bf31-9e8bc453d778" (UID: "4cb15ea4-0788-4003-bf31-9e8bc453d778"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.935457 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4cb15ea4-0788-4003-bf31-9e8bc453d778" (UID: "4cb15ea4-0788-4003-bf31-9e8bc453d778"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.935505 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4cb15ea4-0788-4003-bf31-9e8bc453d778" (UID: "4cb15ea4-0788-4003-bf31-9e8bc453d778"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.935527 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run" (OuterVolumeSpecName: "var-run") pod "4cb15ea4-0788-4003-bf31-9e8bc453d778" (UID: "4cb15ea4-0788-4003-bf31-9e8bc453d778"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.936718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-scripts" (OuterVolumeSpecName: "scripts") pod "4cb15ea4-0788-4003-bf31-9e8bc453d778" (UID: "4cb15ea4-0788-4003-bf31-9e8bc453d778"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:37 crc kubenswrapper[4744]: I1201 08:34:37.950982 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cb15ea4-0788-4003-bf31-9e8bc453d778-kube-api-access-bjltk" (OuterVolumeSpecName: "kube-api-access-bjltk") pod "4cb15ea4-0788-4003-bf31-9e8bc453d778" (UID: "4cb15ea4-0788-4003-bf31-9e8bc453d778"). InnerVolumeSpecName "kube-api-access-bjltk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.036994 4744 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.037032 4744 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.037048 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.037060 4744 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb15ea4-0788-4003-bf31-9e8bc453d778-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.037072 4744 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb15ea4-0788-4003-bf31-9e8bc453d778-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.037085 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjltk\" (UniqueName: \"kubernetes.io/projected/4cb15ea4-0788-4003-bf31-9e8bc453d778-kube-api-access-bjltk\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.309451 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pgp2f-config-m7blz" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.310139 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pgp2f-config-m7blz" event={"ID":"4cb15ea4-0788-4003-bf31-9e8bc453d778","Type":"ContainerDied","Data":"73b8cd8dba92f48e9f8c0e14c695288a5e8441be39586ac89a7f54048057e2b8"} Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.310224 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73b8cd8dba92f48e9f8c0e14c695288a5e8441be39586ac89a7f54048057e2b8" Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.310280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"3ba55e66b75bc120f1eeafa9b98255bf2f97c61cc9d78f640938ea84f7cfd9b8"} Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.310335 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"675af6fbeee57c9b35893744ad5ee87bfa84a58c4d902b4a7971e93882e40305"} Dec 01 08:34:38 crc kubenswrapper[4744]: I1201 08:34:38.993021 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-pgp2f-config-m7blz"] Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.000205 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-pgp2f-config-m7blz"] Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.332208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"cc3bf23e8a7466cd431c7c10dcfade59ca4ccb0c7ee978f0067e5f5b4c56edec"} Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.332246 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"8a2090a929f097d253b6cb8640cb27adc162efe8164a8dbfe364a0401876c47c"} Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.332257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"77aed0d1e5d9e4583d9902fe49c4fd2ef55fb21be190523e10000127b79ad232"} Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.332264 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"903e6d121d71de3ca4baf5eba5fc4154aec8718ef2b059776e2e7b960102ce62"} Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.332273 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"424071bb-0b84-421b-9858-016c60e89b30","Type":"ContainerStarted","Data":"a98d4c37110120fd242e92125c28310fce90dc133ce19d7d1ecccb2bed8fb3f4"} Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.561972 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-pgp2f" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.582998 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.038270096 podStartE2EDuration="28.582979899s" podCreationTimestamp="2025-12-01 08:34:11 +0000 UTC" firstStartedPulling="2025-12-01 08:34:29.259432941 +0000 UTC m=+1001.248490862" lastFinishedPulling="2025-12-01 08:34:37.804142744 +0000 UTC m=+1009.793200665" observedRunningTime="2025-12-01 08:34:39.398345047 +0000 UTC m=+1011.387402968" watchObservedRunningTime="2025-12-01 08:34:39.582979899 +0000 UTC m=+1011.572037820" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.768324 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-7csls"] Dec 01 08:34:39 crc kubenswrapper[4744]: E1201 08:34:39.768657 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb15ea4-0788-4003-bf31-9e8bc453d778" containerName="ovn-config" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.768673 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb15ea4-0788-4003-bf31-9e8bc453d778" containerName="ovn-config" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.768851 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cb15ea4-0788-4003-bf31-9e8bc453d778" containerName="ovn-config" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.769662 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.774071 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.777729 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-7csls"] Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.863895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-config\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.863942 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.863978 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.864329 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.864509 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s6cx\" (UniqueName: \"kubernetes.io/projected/68fd39e9-b7c3-42db-8fcb-a0b39f586693-kube-api-access-7s6cx\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.864549 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.966146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s6cx\" (UniqueName: \"kubernetes.io/projected/68fd39e9-b7c3-42db-8fcb-a0b39f586693-kube-api-access-7s6cx\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.966196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.966249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-config\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.966284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.966341 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.966449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.967454 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.967694 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.967900 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.968045 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-config\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.968450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:39 crc kubenswrapper[4744]: I1201 08:34:39.987453 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s6cx\" (UniqueName: \"kubernetes.io/projected/68fd39e9-b7c3-42db-8fcb-a0b39f586693-kube-api-access-7s6cx\") pod \"dnsmasq-dns-6d5b6d6b67-7csls\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:40 crc kubenswrapper[4744]: I1201 08:34:40.083768 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:40 crc kubenswrapper[4744]: I1201 08:34:40.296595 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cb15ea4-0788-4003-bf31-9e8bc453d778" path="/var/lib/kubelet/pods/4cb15ea4-0788-4003-bf31-9e8bc453d778/volumes" Dec 01 08:34:40 crc kubenswrapper[4744]: I1201 08:34:40.588790 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-7csls"] Dec 01 08:34:40 crc kubenswrapper[4744]: W1201 08:34:40.593960 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68fd39e9_b7c3_42db_8fcb_a0b39f586693.slice/crio-1e8a3f4beb49a113618e2e1b1ea76e0fd75e4db640c6098aa0eaf5abbbedbbc9 WatchSource:0}: Error finding container 1e8a3f4beb49a113618e2e1b1ea76e0fd75e4db640c6098aa0eaf5abbbedbbc9: Status 404 returned error can't find the container with id 1e8a3f4beb49a113618e2e1b1ea76e0fd75e4db640c6098aa0eaf5abbbedbbc9 Dec 01 08:34:41 crc kubenswrapper[4744]: I1201 08:34:41.350812 4744 generic.go:334] "Generic (PLEG): container finished" podID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerID="dcf876e7115e0164cd133ed8d7f5180a7f7826bc99870fff34bf83aec915f4c7" exitCode=0 Dec 01 08:34:41 crc kubenswrapper[4744]: I1201 08:34:41.350897 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" event={"ID":"68fd39e9-b7c3-42db-8fcb-a0b39f586693","Type":"ContainerDied","Data":"dcf876e7115e0164cd133ed8d7f5180a7f7826bc99870fff34bf83aec915f4c7"} Dec 01 08:34:41 crc kubenswrapper[4744]: I1201 08:34:41.351195 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" event={"ID":"68fd39e9-b7c3-42db-8fcb-a0b39f586693","Type":"ContainerStarted","Data":"1e8a3f4beb49a113618e2e1b1ea76e0fd75e4db640c6098aa0eaf5abbbedbbc9"} Dec 01 08:34:42 crc kubenswrapper[4744]: I1201 08:34:42.367524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" event={"ID":"68fd39e9-b7c3-42db-8fcb-a0b39f586693","Type":"ContainerStarted","Data":"fdcc8814fe3a8144d45af81809fe8b51fb45c66b00bff436699248088a1eb92b"} Dec 01 08:34:42 crc kubenswrapper[4744]: I1201 08:34:42.367943 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:42 crc kubenswrapper[4744]: I1201 08:34:42.401851 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" podStartSLOduration=3.40182603 podStartE2EDuration="3.40182603s" podCreationTimestamp="2025-12-01 08:34:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:34:42.389624969 +0000 UTC m=+1014.378682960" watchObservedRunningTime="2025-12-01 08:34:42.40182603 +0000 UTC m=+1014.390883991" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.148637 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.452654 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.592591 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wg8gw"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.593552 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.603965 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6364-account-create-update-9bhwp"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.605030 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.611668 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.617011 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wg8gw"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.633654 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6364-account-create-update-9bhwp"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.684949 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-9tggf"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.686036 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.693213 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-9tggf"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.707330 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e015-account-create-update-m75kd"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.710261 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.714104 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.742837 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e015-account-create-update-m75kd"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.762045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptm8d\" (UniqueName: \"kubernetes.io/projected/8ac6bc34-6d3b-4860-aab8-bdf00a485793-kube-api-access-ptm8d\") pod \"cinder-db-create-wg8gw\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.762094 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ac6bc34-6d3b-4860-aab8-bdf00a485793-operator-scripts\") pod \"cinder-db-create-wg8gw\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.762147 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4341bae7-8d88-4d04-8670-43125be4a7a3-operator-scripts\") pod \"cinder-6364-account-create-update-9bhwp\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.762191 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptsbg\" (UniqueName: \"kubernetes.io/projected/4341bae7-8d88-4d04-8670-43125be4a7a3-kube-api-access-ptsbg\") pod \"cinder-6364-account-create-update-9bhwp\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864072 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqg5h\" (UniqueName: \"kubernetes.io/projected/ed9aa660-b80d-4dd2-a150-fee9352d72b8-kube-api-access-wqg5h\") pod \"barbican-e015-account-create-update-m75kd\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864153 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptm8d\" (UniqueName: \"kubernetes.io/projected/8ac6bc34-6d3b-4860-aab8-bdf00a485793-kube-api-access-ptm8d\") pod \"cinder-db-create-wg8gw\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864183 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ac6bc34-6d3b-4860-aab8-bdf00a485793-operator-scripts\") pod \"cinder-db-create-wg8gw\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864206 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b116a0ec-0413-4532-b820-181cbf9f3515-operator-scripts\") pod \"barbican-db-create-9tggf\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864253 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9aa660-b80d-4dd2-a150-fee9352d72b8-operator-scripts\") pod \"barbican-e015-account-create-update-m75kd\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864274 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4341bae7-8d88-4d04-8670-43125be4a7a3-operator-scripts\") pod \"cinder-6364-account-create-update-9bhwp\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864317 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llf2x\" (UniqueName: \"kubernetes.io/projected/b116a0ec-0413-4532-b820-181cbf9f3515-kube-api-access-llf2x\") pod \"barbican-db-create-9tggf\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.864334 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptsbg\" (UniqueName: \"kubernetes.io/projected/4341bae7-8d88-4d04-8670-43125be4a7a3-kube-api-access-ptsbg\") pod \"cinder-6364-account-create-update-9bhwp\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.865326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ac6bc34-6d3b-4860-aab8-bdf00a485793-operator-scripts\") pod \"cinder-db-create-wg8gw\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.865843 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4341bae7-8d88-4d04-8670-43125be4a7a3-operator-scripts\") pod \"cinder-6364-account-create-update-9bhwp\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.879368 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-gx62t"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.882718 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gx62t" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.891184 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptm8d\" (UniqueName: \"kubernetes.io/projected/8ac6bc34-6d3b-4860-aab8-bdf00a485793-kube-api-access-ptm8d\") pod \"cinder-db-create-wg8gw\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.898558 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-b269-account-create-update-5rlsq"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.899586 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.901111 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.910180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptsbg\" (UniqueName: \"kubernetes.io/projected/4341bae7-8d88-4d04-8670-43125be4a7a3-kube-api-access-ptsbg\") pod \"cinder-6364-account-create-update-9bhwp\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.919013 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.921382 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-gx62t"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.929130 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.962723 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-b269-account-create-update-5rlsq"] Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.966649 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqg5h\" (UniqueName: \"kubernetes.io/projected/ed9aa660-b80d-4dd2-a150-fee9352d72b8-kube-api-access-wqg5h\") pod \"barbican-e015-account-create-update-m75kd\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.966723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s8cj\" (UniqueName: \"kubernetes.io/projected/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-kube-api-access-9s8cj\") pod \"heat-db-create-gx62t\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " pod="openstack/heat-db-create-gx62t" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.966781 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b116a0ec-0413-4532-b820-181cbf9f3515-operator-scripts\") pod \"barbican-db-create-9tggf\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.966830 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9aa660-b80d-4dd2-a150-fee9352d72b8-operator-scripts\") pod \"barbican-e015-account-create-update-m75kd\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.966866 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-operator-scripts\") pod \"heat-db-create-gx62t\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " pod="openstack/heat-db-create-gx62t" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.966901 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llf2x\" (UniqueName: \"kubernetes.io/projected/b116a0ec-0413-4532-b820-181cbf9f3515-kube-api-access-llf2x\") pod \"barbican-db-create-9tggf\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.972883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9aa660-b80d-4dd2-a150-fee9352d72b8-operator-scripts\") pod \"barbican-e015-account-create-update-m75kd\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.978484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b116a0ec-0413-4532-b820-181cbf9f3515-operator-scripts\") pod \"barbican-db-create-9tggf\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:45 crc kubenswrapper[4744]: I1201 08:34:45.987527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llf2x\" (UniqueName: \"kubernetes.io/projected/b116a0ec-0413-4532-b820-181cbf9f3515-kube-api-access-llf2x\") pod \"barbican-db-create-9tggf\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.001815 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.004683 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqg5h\" (UniqueName: \"kubernetes.io/projected/ed9aa660-b80d-4dd2-a150-fee9352d72b8-kube-api-access-wqg5h\") pod \"barbican-e015-account-create-update-m75kd\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.009336 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-d42bc"] Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.012367 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.016068 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.016321 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.016358 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.031628 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4jgn6" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.031796 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.048156 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-d42bc"] Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.068081 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6sxv\" (UniqueName: \"kubernetes.io/projected/6e4d4551-0e28-432e-bd4e-e783133b0a4b-kube-api-access-k6sxv\") pod \"heat-b269-account-create-update-5rlsq\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.068125 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s8cj\" (UniqueName: \"kubernetes.io/projected/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-kube-api-access-9s8cj\") pod \"heat-db-create-gx62t\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " pod="openstack/heat-db-create-gx62t" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.068240 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-operator-scripts\") pod \"heat-db-create-gx62t\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " pod="openstack/heat-db-create-gx62t" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.068288 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e4d4551-0e28-432e-bd4e-e783133b0a4b-operator-scripts\") pod \"heat-b269-account-create-update-5rlsq\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.072809 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-operator-scripts\") pod \"heat-db-create-gx62t\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " pod="openstack/heat-db-create-gx62t" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.089811 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s8cj\" (UniqueName: \"kubernetes.io/projected/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-kube-api-access-9s8cj\") pod \"heat-db-create-gx62t\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " pod="openstack/heat-db-create-gx62t" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.101462 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-fa0a-account-create-update-stqk2"] Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.102650 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.104673 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.143691 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fa0a-account-create-update-stqk2"] Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.170203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-combined-ca-bundle\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.170320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e4d4551-0e28-432e-bd4e-e783133b0a4b-operator-scripts\") pod \"heat-b269-account-create-update-5rlsq\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.170341 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggbpl\" (UniqueName: \"kubernetes.io/projected/d725b87c-b27b-414d-824d-e9f9f3172d69-kube-api-access-ggbpl\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.170368 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6sxv\" (UniqueName: \"kubernetes.io/projected/6e4d4551-0e28-432e-bd4e-e783133b0a4b-kube-api-access-k6sxv\") pod \"heat-b269-account-create-update-5rlsq\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.170435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-config-data\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.171571 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e4d4551-0e28-432e-bd4e-e783133b0a4b-operator-scripts\") pod \"heat-b269-account-create-update-5rlsq\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.178901 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-4xvfq"] Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.180034 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.191058 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6sxv\" (UniqueName: \"kubernetes.io/projected/6e4d4551-0e28-432e-bd4e-e783133b0a4b-kube-api-access-k6sxv\") pod \"heat-b269-account-create-update-5rlsq\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.191878 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4xvfq"] Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.271259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-config-data\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.271302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-combined-ca-bundle\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.271362 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/198a3f7b-ed68-4dee-92e1-9aa0e9362304-operator-scripts\") pod \"neutron-fa0a-account-create-update-stqk2\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.271962 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5076354c-7f79-490f-9ea7-c5d2dd60c545-operator-scripts\") pod \"neutron-db-create-4xvfq\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.272074 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkrpm\" (UniqueName: \"kubernetes.io/projected/5076354c-7f79-490f-9ea7-c5d2dd60c545-kube-api-access-jkrpm\") pod \"neutron-db-create-4xvfq\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.272105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggbpl\" (UniqueName: \"kubernetes.io/projected/d725b87c-b27b-414d-824d-e9f9f3172d69-kube-api-access-ggbpl\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.272147 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2bbt\" (UniqueName: \"kubernetes.io/projected/198a3f7b-ed68-4dee-92e1-9aa0e9362304-kube-api-access-z2bbt\") pod \"neutron-fa0a-account-create-update-stqk2\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.274924 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-config-data\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.275351 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-combined-ca-bundle\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.293192 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggbpl\" (UniqueName: \"kubernetes.io/projected/d725b87c-b27b-414d-824d-e9f9f3172d69-kube-api-access-ggbpl\") pod \"keystone-db-sync-d42bc\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.374193 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5076354c-7f79-490f-9ea7-c5d2dd60c545-operator-scripts\") pod \"neutron-db-create-4xvfq\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.374255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkrpm\" (UniqueName: \"kubernetes.io/projected/5076354c-7f79-490f-9ea7-c5d2dd60c545-kube-api-access-jkrpm\") pod \"neutron-db-create-4xvfq\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.374349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2bbt\" (UniqueName: \"kubernetes.io/projected/198a3f7b-ed68-4dee-92e1-9aa0e9362304-kube-api-access-z2bbt\") pod \"neutron-fa0a-account-create-update-stqk2\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.375108 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5076354c-7f79-490f-9ea7-c5d2dd60c545-operator-scripts\") pod \"neutron-db-create-4xvfq\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.375537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/198a3f7b-ed68-4dee-92e1-9aa0e9362304-operator-scripts\") pod \"neutron-fa0a-account-create-update-stqk2\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.375857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/198a3f7b-ed68-4dee-92e1-9aa0e9362304-operator-scripts\") pod \"neutron-fa0a-account-create-update-stqk2\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.389985 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gx62t" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.396514 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2bbt\" (UniqueName: \"kubernetes.io/projected/198a3f7b-ed68-4dee-92e1-9aa0e9362304-kube-api-access-z2bbt\") pod \"neutron-fa0a-account-create-update-stqk2\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.397027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkrpm\" (UniqueName: \"kubernetes.io/projected/5076354c-7f79-490f-9ea7-c5d2dd60c545-kube-api-access-jkrpm\") pod \"neutron-db-create-4xvfq\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.403529 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.415932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d42bc" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.443758 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.505867 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.557067 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e015-account-create-update-m75kd"] Dec 01 08:34:46 crc kubenswrapper[4744]: W1201 08:34:46.574230 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded9aa660_b80d_4dd2_a150_fee9352d72b8.slice/crio-c8689fe53b6d835cb7d594850c99cc970b893f98dc80bad327aa74a59197c54d WatchSource:0}: Error finding container c8689fe53b6d835cb7d594850c99cc970b893f98dc80bad327aa74a59197c54d: Status 404 returned error can't find the container with id c8689fe53b6d835cb7d594850c99cc970b893f98dc80bad327aa74a59197c54d Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.632914 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6364-account-create-update-9bhwp"] Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.663587 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wg8gw"] Dec 01 08:34:46 crc kubenswrapper[4744]: W1201 08:34:46.676749 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ac6bc34_6d3b_4860_aab8_bdf00a485793.slice/crio-9e58746ad8b109fde6dc85198ab14b3c59e2974037a1d2e624ea7aaf71f720e4 WatchSource:0}: Error finding container 9e58746ad8b109fde6dc85198ab14b3c59e2974037a1d2e624ea7aaf71f720e4: Status 404 returned error can't find the container with id 9e58746ad8b109fde6dc85198ab14b3c59e2974037a1d2e624ea7aaf71f720e4 Dec 01 08:34:46 crc kubenswrapper[4744]: I1201 08:34:46.679962 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-9tggf"] Dec 01 08:34:46 crc kubenswrapper[4744]: W1201 08:34:46.702747 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb116a0ec_0413_4532_b820_181cbf9f3515.slice/crio-0ebedd145a83e98008129e407e8b33e0162b0eba02b4f9c2664d5ea3b4d3ec1b WatchSource:0}: Error finding container 0ebedd145a83e98008129e407e8b33e0162b0eba02b4f9c2664d5ea3b4d3ec1b: Status 404 returned error can't find the container with id 0ebedd145a83e98008129e407e8b33e0162b0eba02b4f9c2664d5ea3b4d3ec1b Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.069308 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-gx62t"] Dec 01 08:34:47 crc kubenswrapper[4744]: W1201 08:34:47.069377 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fb2f8c0_0676_4f02_af2c_b4000c09ab26.slice/crio-1815864c6329e3dcf1c017ea95b4f5590b66f712df97675275c58ddb254c0413 WatchSource:0}: Error finding container 1815864c6329e3dcf1c017ea95b4f5590b66f712df97675275c58ddb254c0413: Status 404 returned error can't find the container with id 1815864c6329e3dcf1c017ea95b4f5590b66f712df97675275c58ddb254c0413 Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.142049 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fa0a-account-create-update-stqk2"] Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.162567 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4xvfq"] Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.193373 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-b269-account-create-update-5rlsq"] Dec 01 08:34:47 crc kubenswrapper[4744]: W1201 08:34:47.228258 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e4d4551_0e28_432e_bd4e_e783133b0a4b.slice/crio-acf6ea1463b1cd45a0d396da2a49c96d1d1dde87cf073c93d3a717ca69d675a0 WatchSource:0}: Error finding container acf6ea1463b1cd45a0d396da2a49c96d1d1dde87cf073c93d3a717ca69d675a0: Status 404 returned error can't find the container with id acf6ea1463b1cd45a0d396da2a49c96d1d1dde87cf073c93d3a717ca69d675a0 Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.269519 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-d42bc"] Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.479909 4744 generic.go:334] "Generic (PLEG): container finished" podID="4341bae7-8d88-4d04-8670-43125be4a7a3" containerID="6f5660f26c5029a43db70500fcfb1dfeb7ad53b0b1d43841e145bf009afb4e06" exitCode=0 Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.479989 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6364-account-create-update-9bhwp" event={"ID":"4341bae7-8d88-4d04-8670-43125be4a7a3","Type":"ContainerDied","Data":"6f5660f26c5029a43db70500fcfb1dfeb7ad53b0b1d43841e145bf009afb4e06"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.480026 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6364-account-create-update-9bhwp" event={"ID":"4341bae7-8d88-4d04-8670-43125be4a7a3","Type":"ContainerStarted","Data":"ee86bc1f4ce7b3a93e023d9d846b7c9f7ca23a59e20d3b3bba690aedd99ca63e"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.482802 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4xvfq" event={"ID":"5076354c-7f79-490f-9ea7-c5d2dd60c545","Type":"ContainerStarted","Data":"c1188526a4dfe35bc048bc4cc0dcb13ab689830a290a13df9f469f2b15e58853"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.485785 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fa0a-account-create-update-stqk2" event={"ID":"198a3f7b-ed68-4dee-92e1-9aa0e9362304","Type":"ContainerStarted","Data":"9bb469792f3932c861d4dddd7cab292baf01e1152bb8aecceedd6dbf55cf187a"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.488120 4744 generic.go:334] "Generic (PLEG): container finished" podID="ed9aa660-b80d-4dd2-a150-fee9352d72b8" containerID="46888f197156e6ef7c11d27d741085d1867e097fda7e53af0756b6301e4a36c4" exitCode=0 Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.488192 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e015-account-create-update-m75kd" event={"ID":"ed9aa660-b80d-4dd2-a150-fee9352d72b8","Type":"ContainerDied","Data":"46888f197156e6ef7c11d27d741085d1867e097fda7e53af0756b6301e4a36c4"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.488218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e015-account-create-update-m75kd" event={"ID":"ed9aa660-b80d-4dd2-a150-fee9352d72b8","Type":"ContainerStarted","Data":"c8689fe53b6d835cb7d594850c99cc970b893f98dc80bad327aa74a59197c54d"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.490967 4744 generic.go:334] "Generic (PLEG): container finished" podID="b116a0ec-0413-4532-b820-181cbf9f3515" containerID="982dfc2415048df980ec9f5d149660f2d3691dc57ad83c38d7fff673b2fea9b8" exitCode=0 Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.491019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9tggf" event={"ID":"b116a0ec-0413-4532-b820-181cbf9f3515","Type":"ContainerDied","Data":"982dfc2415048df980ec9f5d149660f2d3691dc57ad83c38d7fff673b2fea9b8"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.491038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9tggf" event={"ID":"b116a0ec-0413-4532-b820-181cbf9f3515","Type":"ContainerStarted","Data":"0ebedd145a83e98008129e407e8b33e0162b0eba02b4f9c2664d5ea3b4d3ec1b"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.492484 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ac6bc34-6d3b-4860-aab8-bdf00a485793" containerID="44b0a76ff59530d05f1988f9403e2f890dd600e1d7a43a60e5cb7ffeff13911c" exitCode=0 Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.493286 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wg8gw" event={"ID":"8ac6bc34-6d3b-4860-aab8-bdf00a485793","Type":"ContainerDied","Data":"44b0a76ff59530d05f1988f9403e2f890dd600e1d7a43a60e5cb7ffeff13911c"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.493372 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wg8gw" event={"ID":"8ac6bc34-6d3b-4860-aab8-bdf00a485793","Type":"ContainerStarted","Data":"9e58746ad8b109fde6dc85198ab14b3c59e2974037a1d2e624ea7aaf71f720e4"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.499645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gx62t" event={"ID":"5fb2f8c0-0676-4f02-af2c-b4000c09ab26","Type":"ContainerStarted","Data":"1815864c6329e3dcf1c017ea95b4f5590b66f712df97675275c58ddb254c0413"} Dec 01 08:34:47 crc kubenswrapper[4744]: I1201 08:34:47.506836 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b269-account-create-update-5rlsq" event={"ID":"6e4d4551-0e28-432e-bd4e-e783133b0a4b","Type":"ContainerStarted","Data":"acf6ea1463b1cd45a0d396da2a49c96d1d1dde87cf073c93d3a717ca69d675a0"} Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.576050 4744 generic.go:334] "Generic (PLEG): container finished" podID="6e4d4551-0e28-432e-bd4e-e783133b0a4b" containerID="699bb5033f2dcaae20c9ce9cfa7d30a9bafa0539eb2d6ca6443120d168c80d6a" exitCode=0 Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.576212 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b269-account-create-update-5rlsq" event={"ID":"6e4d4551-0e28-432e-bd4e-e783133b0a4b","Type":"ContainerDied","Data":"699bb5033f2dcaae20c9ce9cfa7d30a9bafa0539eb2d6ca6443120d168c80d6a"} Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.606710 4744 generic.go:334] "Generic (PLEG): container finished" podID="5076354c-7f79-490f-9ea7-c5d2dd60c545" containerID="bc239bc0d3c935b996af080f41b06f11a86f65fabe629577114ed79e0af492b0" exitCode=0 Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.606835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4xvfq" event={"ID":"5076354c-7f79-490f-9ea7-c5d2dd60c545","Type":"ContainerDied","Data":"bc239bc0d3c935b996af080f41b06f11a86f65fabe629577114ed79e0af492b0"} Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.608561 4744 generic.go:334] "Generic (PLEG): container finished" podID="198a3f7b-ed68-4dee-92e1-9aa0e9362304" containerID="7895b27b5816872126dcec77e8860c51572c626aad714162aca663072a7a3df7" exitCode=0 Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.608633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fa0a-account-create-update-stqk2" event={"ID":"198a3f7b-ed68-4dee-92e1-9aa0e9362304","Type":"ContainerDied","Data":"7895b27b5816872126dcec77e8860c51572c626aad714162aca663072a7a3df7"} Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.637557 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.637615 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.637659 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.637942 4744 generic.go:334] "Generic (PLEG): container finished" podID="5fb2f8c0-0676-4f02-af2c-b4000c09ab26" containerID="032b78dc265273587187f76090f5e446c42b62225f02ffb574c36c8b8411595a" exitCode=0 Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.638008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gx62t" event={"ID":"5fb2f8c0-0676-4f02-af2c-b4000c09ab26","Type":"ContainerDied","Data":"032b78dc265273587187f76090f5e446c42b62225f02ffb574c36c8b8411595a"} Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.638289 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64ec6b8815aa74952cf65128020a2c3112aeb4c1255a0d0fff478a4f4a871f94"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.638331 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://64ec6b8815aa74952cf65128020a2c3112aeb4c1255a0d0fff478a4f4a871f94" gracePeriod=600 Dec 01 08:34:48 crc kubenswrapper[4744]: I1201 08:34:48.644305 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d42bc" event={"ID":"d725b87c-b27b-414d-824d-e9f9f3172d69","Type":"ContainerStarted","Data":"c633121d94932e5597e768674e81acccaa39c466112e570daa965193ec53a20d"} Dec 01 08:34:49 crc kubenswrapper[4744]: I1201 08:34:49.660525 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="64ec6b8815aa74952cf65128020a2c3112aeb4c1255a0d0fff478a4f4a871f94" exitCode=0 Dec 01 08:34:49 crc kubenswrapper[4744]: I1201 08:34:49.660579 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"64ec6b8815aa74952cf65128020a2c3112aeb4c1255a0d0fff478a4f4a871f94"} Dec 01 08:34:49 crc kubenswrapper[4744]: I1201 08:34:49.660645 4744 scope.go:117] "RemoveContainer" containerID="0d6c923f3edaf0f41eb1f1e74b19db369be4219783160ca17cb5ceacdf850175" Dec 01 08:34:50 crc kubenswrapper[4744]: I1201 08:34:50.086605 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:34:50 crc kubenswrapper[4744]: I1201 08:34:50.141982 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fbvs9"] Dec 01 08:34:50 crc kubenswrapper[4744]: I1201 08:34:50.142246 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" podUID="3154b059-5545-416d-aca9-426f493b85f2" containerName="dnsmasq-dns" containerID="cri-o://6e14215e6365c84aca27e254ef9aca2511a6650ff02052805683526f016d096c" gracePeriod=10 Dec 01 08:34:50 crc kubenswrapper[4744]: I1201 08:34:50.673328 4744 generic.go:334] "Generic (PLEG): container finished" podID="3154b059-5545-416d-aca9-426f493b85f2" containerID="6e14215e6365c84aca27e254ef9aca2511a6650ff02052805683526f016d096c" exitCode=0 Dec 01 08:34:50 crc kubenswrapper[4744]: I1201 08:34:50.673378 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" event={"ID":"3154b059-5545-416d-aca9-426f493b85f2","Type":"ContainerDied","Data":"6e14215e6365c84aca27e254ef9aca2511a6650ff02052805683526f016d096c"} Dec 01 08:34:51 crc kubenswrapper[4744]: I1201 08:34:51.534743 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" podUID="3154b059-5545-416d-aca9-426f493b85f2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.288843 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.304802 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.344964 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.370432 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.375071 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.409481 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.419978 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.431399 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gx62t" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.436558 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.475697 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkrpm\" (UniqueName: \"kubernetes.io/projected/5076354c-7f79-490f-9ea7-c5d2dd60c545-kube-api-access-jkrpm\") pod \"5076354c-7f79-490f-9ea7-c5d2dd60c545\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.475749 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llf2x\" (UniqueName: \"kubernetes.io/projected/b116a0ec-0413-4532-b820-181cbf9f3515-kube-api-access-llf2x\") pod \"b116a0ec-0413-4532-b820-181cbf9f3515\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.475768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e4d4551-0e28-432e-bd4e-e783133b0a4b-operator-scripts\") pod \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.475822 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6sxv\" (UniqueName: \"kubernetes.io/projected/6e4d4551-0e28-432e-bd4e-e783133b0a4b-kube-api-access-k6sxv\") pod \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\" (UID: \"6e4d4551-0e28-432e-bd4e-e783133b0a4b\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.475877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b116a0ec-0413-4532-b820-181cbf9f3515-operator-scripts\") pod \"b116a0ec-0413-4532-b820-181cbf9f3515\" (UID: \"b116a0ec-0413-4532-b820-181cbf9f3515\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.475907 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9aa660-b80d-4dd2-a150-fee9352d72b8-operator-scripts\") pod \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.475952 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5076354c-7f79-490f-9ea7-c5d2dd60c545-operator-scripts\") pod \"5076354c-7f79-490f-9ea7-c5d2dd60c545\" (UID: \"5076354c-7f79-490f-9ea7-c5d2dd60c545\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.476026 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ac6bc34-6d3b-4860-aab8-bdf00a485793-operator-scripts\") pod \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.476049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptm8d\" (UniqueName: \"kubernetes.io/projected/8ac6bc34-6d3b-4860-aab8-bdf00a485793-kube-api-access-ptm8d\") pod \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\" (UID: \"8ac6bc34-6d3b-4860-aab8-bdf00a485793\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.476158 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqg5h\" (UniqueName: \"kubernetes.io/projected/ed9aa660-b80d-4dd2-a150-fee9352d72b8-kube-api-access-wqg5h\") pod \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\" (UID: \"ed9aa660-b80d-4dd2-a150-fee9352d72b8\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.476511 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b116a0ec-0413-4532-b820-181cbf9f3515-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b116a0ec-0413-4532-b820-181cbf9f3515" (UID: "b116a0ec-0413-4532-b820-181cbf9f3515"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.476654 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b116a0ec-0413-4532-b820-181cbf9f3515-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.477032 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e4d4551-0e28-432e-bd4e-e783133b0a4b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6e4d4551-0e28-432e-bd4e-e783133b0a4b" (UID: "6e4d4551-0e28-432e-bd4e-e783133b0a4b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.477149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9aa660-b80d-4dd2-a150-fee9352d72b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed9aa660-b80d-4dd2-a150-fee9352d72b8" (UID: "ed9aa660-b80d-4dd2-a150-fee9352d72b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.480995 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b116a0ec-0413-4532-b820-181cbf9f3515-kube-api-access-llf2x" (OuterVolumeSpecName: "kube-api-access-llf2x") pod "b116a0ec-0413-4532-b820-181cbf9f3515" (UID: "b116a0ec-0413-4532-b820-181cbf9f3515"). InnerVolumeSpecName "kube-api-access-llf2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.482073 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ac6bc34-6d3b-4860-aab8-bdf00a485793-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8ac6bc34-6d3b-4860-aab8-bdf00a485793" (UID: "8ac6bc34-6d3b-4860-aab8-bdf00a485793"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.482270 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5076354c-7f79-490f-9ea7-c5d2dd60c545-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5076354c-7f79-490f-9ea7-c5d2dd60c545" (UID: "5076354c-7f79-490f-9ea7-c5d2dd60c545"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.482298 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9aa660-b80d-4dd2-a150-fee9352d72b8-kube-api-access-wqg5h" (OuterVolumeSpecName: "kube-api-access-wqg5h") pod "ed9aa660-b80d-4dd2-a150-fee9352d72b8" (UID: "ed9aa660-b80d-4dd2-a150-fee9352d72b8"). InnerVolumeSpecName "kube-api-access-wqg5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.483291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e4d4551-0e28-432e-bd4e-e783133b0a4b-kube-api-access-k6sxv" (OuterVolumeSpecName: "kube-api-access-k6sxv") pod "6e4d4551-0e28-432e-bd4e-e783133b0a4b" (UID: "6e4d4551-0e28-432e-bd4e-e783133b0a4b"). InnerVolumeSpecName "kube-api-access-k6sxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.485169 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5076354c-7f79-490f-9ea7-c5d2dd60c545-kube-api-access-jkrpm" (OuterVolumeSpecName: "kube-api-access-jkrpm") pod "5076354c-7f79-490f-9ea7-c5d2dd60c545" (UID: "5076354c-7f79-490f-9ea7-c5d2dd60c545"). InnerVolumeSpecName "kube-api-access-jkrpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.486072 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ac6bc34-6d3b-4860-aab8-bdf00a485793-kube-api-access-ptm8d" (OuterVolumeSpecName: "kube-api-access-ptm8d") pod "8ac6bc34-6d3b-4860-aab8-bdf00a485793" (UID: "8ac6bc34-6d3b-4860-aab8-bdf00a485793"). InnerVolumeSpecName "kube-api-access-ptm8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.577990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfxs5\" (UniqueName: \"kubernetes.io/projected/3154b059-5545-416d-aca9-426f493b85f2-kube-api-access-gfxs5\") pod \"3154b059-5545-416d-aca9-426f493b85f2\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578359 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2bbt\" (UniqueName: \"kubernetes.io/projected/198a3f7b-ed68-4dee-92e1-9aa0e9362304-kube-api-access-z2bbt\") pod \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578444 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-dns-svc\") pod \"3154b059-5545-416d-aca9-426f493b85f2\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578491 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-sb\") pod \"3154b059-5545-416d-aca9-426f493b85f2\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578538 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-operator-scripts\") pod \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578648 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptsbg\" (UniqueName: \"kubernetes.io/projected/4341bae7-8d88-4d04-8670-43125be4a7a3-kube-api-access-ptsbg\") pod \"4341bae7-8d88-4d04-8670-43125be4a7a3\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578741 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/198a3f7b-ed68-4dee-92e1-9aa0e9362304-operator-scripts\") pod \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\" (UID: \"198a3f7b-ed68-4dee-92e1-9aa0e9362304\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578771 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-config\") pod \"3154b059-5545-416d-aca9-426f493b85f2\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578807 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4341bae7-8d88-4d04-8670-43125be4a7a3-operator-scripts\") pod \"4341bae7-8d88-4d04-8670-43125be4a7a3\" (UID: \"4341bae7-8d88-4d04-8670-43125be4a7a3\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578878 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s8cj\" (UniqueName: \"kubernetes.io/projected/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-kube-api-access-9s8cj\") pod \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\" (UID: \"5fb2f8c0-0676-4f02-af2c-b4000c09ab26\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.578899 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-nb\") pod \"3154b059-5545-416d-aca9-426f493b85f2\" (UID: \"3154b059-5545-416d-aca9-426f493b85f2\") " Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579286 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkrpm\" (UniqueName: \"kubernetes.io/projected/5076354c-7f79-490f-9ea7-c5d2dd60c545-kube-api-access-jkrpm\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579303 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llf2x\" (UniqueName: \"kubernetes.io/projected/b116a0ec-0413-4532-b820-181cbf9f3515-kube-api-access-llf2x\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579313 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e4d4551-0e28-432e-bd4e-e783133b0a4b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579346 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6sxv\" (UniqueName: \"kubernetes.io/projected/6e4d4551-0e28-432e-bd4e-e783133b0a4b-kube-api-access-k6sxv\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579356 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9aa660-b80d-4dd2-a150-fee9352d72b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579364 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5076354c-7f79-490f-9ea7-c5d2dd60c545-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579374 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ac6bc34-6d3b-4860-aab8-bdf00a485793-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579398 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptm8d\" (UniqueName: \"kubernetes.io/projected/8ac6bc34-6d3b-4860-aab8-bdf00a485793-kube-api-access-ptm8d\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.579421 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqg5h\" (UniqueName: \"kubernetes.io/projected/ed9aa660-b80d-4dd2-a150-fee9352d72b8-kube-api-access-wqg5h\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.588083 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5fb2f8c0-0676-4f02-af2c-b4000c09ab26" (UID: "5fb2f8c0-0676-4f02-af2c-b4000c09ab26"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.592149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3154b059-5545-416d-aca9-426f493b85f2-kube-api-access-gfxs5" (OuterVolumeSpecName: "kube-api-access-gfxs5") pod "3154b059-5545-416d-aca9-426f493b85f2" (UID: "3154b059-5545-416d-aca9-426f493b85f2"). InnerVolumeSpecName "kube-api-access-gfxs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.592193 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4341bae7-8d88-4d04-8670-43125be4a7a3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4341bae7-8d88-4d04-8670-43125be4a7a3" (UID: "4341bae7-8d88-4d04-8670-43125be4a7a3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.592251 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/198a3f7b-ed68-4dee-92e1-9aa0e9362304-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "198a3f7b-ed68-4dee-92e1-9aa0e9362304" (UID: "198a3f7b-ed68-4dee-92e1-9aa0e9362304"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.594284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/198a3f7b-ed68-4dee-92e1-9aa0e9362304-kube-api-access-z2bbt" (OuterVolumeSpecName: "kube-api-access-z2bbt") pod "198a3f7b-ed68-4dee-92e1-9aa0e9362304" (UID: "198a3f7b-ed68-4dee-92e1-9aa0e9362304"). InnerVolumeSpecName "kube-api-access-z2bbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.595560 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4341bae7-8d88-4d04-8670-43125be4a7a3-kube-api-access-ptsbg" (OuterVolumeSpecName: "kube-api-access-ptsbg") pod "4341bae7-8d88-4d04-8670-43125be4a7a3" (UID: "4341bae7-8d88-4d04-8670-43125be4a7a3"). InnerVolumeSpecName "kube-api-access-ptsbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.597600 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-kube-api-access-9s8cj" (OuterVolumeSpecName: "kube-api-access-9s8cj") pod "5fb2f8c0-0676-4f02-af2c-b4000c09ab26" (UID: "5fb2f8c0-0676-4f02-af2c-b4000c09ab26"). InnerVolumeSpecName "kube-api-access-9s8cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.631679 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3154b059-5545-416d-aca9-426f493b85f2" (UID: "3154b059-5545-416d-aca9-426f493b85f2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.641015 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3154b059-5545-416d-aca9-426f493b85f2" (UID: "3154b059-5545-416d-aca9-426f493b85f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.641749 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3154b059-5545-416d-aca9-426f493b85f2" (UID: "3154b059-5545-416d-aca9-426f493b85f2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.648856 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-config" (OuterVolumeSpecName: "config") pod "3154b059-5545-416d-aca9-426f493b85f2" (UID: "3154b059-5545-416d-aca9-426f493b85f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682050 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s8cj\" (UniqueName: \"kubernetes.io/projected/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-kube-api-access-9s8cj\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682089 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682100 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfxs5\" (UniqueName: \"kubernetes.io/projected/3154b059-5545-416d-aca9-426f493b85f2-kube-api-access-gfxs5\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682108 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2bbt\" (UniqueName: \"kubernetes.io/projected/198a3f7b-ed68-4dee-92e1-9aa0e9362304-kube-api-access-z2bbt\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682118 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682126 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682134 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fb2f8c0-0676-4f02-af2c-b4000c09ab26-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682142 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptsbg\" (UniqueName: \"kubernetes.io/projected/4341bae7-8d88-4d04-8670-43125be4a7a3-kube-api-access-ptsbg\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682149 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/198a3f7b-ed68-4dee-92e1-9aa0e9362304-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682157 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3154b059-5545-416d-aca9-426f493b85f2-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.682167 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4341bae7-8d88-4d04-8670-43125be4a7a3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.730883 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4xvfq" event={"ID":"5076354c-7f79-490f-9ea7-c5d2dd60c545","Type":"ContainerDied","Data":"c1188526a4dfe35bc048bc4cc0dcb13ab689830a290a13df9f469f2b15e58853"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.730904 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4xvfq" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.730921 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1188526a4dfe35bc048bc4cc0dcb13ab689830a290a13df9f469f2b15e58853" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.732949 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.732947 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-fbvs9" event={"ID":"3154b059-5545-416d-aca9-426f493b85f2","Type":"ContainerDied","Data":"454152fce33070d2d87453d309b93724f9dde744b8008de3f23989a7160dd33e"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.733060 4744 scope.go:117] "RemoveContainer" containerID="6e14215e6365c84aca27e254ef9aca2511a6650ff02052805683526f016d096c" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.735972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-9tggf" event={"ID":"b116a0ec-0413-4532-b820-181cbf9f3515","Type":"ContainerDied","Data":"0ebedd145a83e98008129e407e8b33e0162b0eba02b4f9c2664d5ea3b4d3ec1b"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.736031 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ebedd145a83e98008129e407e8b33e0162b0eba02b4f9c2664d5ea3b4d3ec1b" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.735987 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-9tggf" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.738877 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wg8gw" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.738877 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wg8gw" event={"ID":"8ac6bc34-6d3b-4860-aab8-bdf00a485793","Type":"ContainerDied","Data":"9e58746ad8b109fde6dc85198ab14b3c59e2974037a1d2e624ea7aaf71f720e4"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.738999 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e58746ad8b109fde6dc85198ab14b3c59e2974037a1d2e624ea7aaf71f720e4" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.740291 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gx62t" event={"ID":"5fb2f8c0-0676-4f02-af2c-b4000c09ab26","Type":"ContainerDied","Data":"1815864c6329e3dcf1c017ea95b4f5590b66f712df97675275c58ddb254c0413"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.740314 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1815864c6329e3dcf1c017ea95b4f5590b66f712df97675275c58ddb254c0413" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.740314 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gx62t" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.746435 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b269-account-create-update-5rlsq" event={"ID":"6e4d4551-0e28-432e-bd4e-e783133b0a4b","Type":"ContainerDied","Data":"acf6ea1463b1cd45a0d396da2a49c96d1d1dde87cf073c93d3a717ca69d675a0"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.746471 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acf6ea1463b1cd45a0d396da2a49c96d1d1dde87cf073c93d3a717ca69d675a0" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.746529 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b269-account-create-update-5rlsq" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.749023 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6364-account-create-update-9bhwp" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.749024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6364-account-create-update-9bhwp" event={"ID":"4341bae7-8d88-4d04-8670-43125be4a7a3","Type":"ContainerDied","Data":"ee86bc1f4ce7b3a93e023d9d846b7c9f7ca23a59e20d3b3bba690aedd99ca63e"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.749176 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee86bc1f4ce7b3a93e023d9d846b7c9f7ca23a59e20d3b3bba690aedd99ca63e" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.750441 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fa0a-account-create-update-stqk2" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.750472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fa0a-account-create-update-stqk2" event={"ID":"198a3f7b-ed68-4dee-92e1-9aa0e9362304","Type":"ContainerDied","Data":"9bb469792f3932c861d4dddd7cab292baf01e1152bb8aecceedd6dbf55cf187a"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.750495 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bb469792f3932c861d4dddd7cab292baf01e1152bb8aecceedd6dbf55cf187a" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.751729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e015-account-create-update-m75kd" event={"ID":"ed9aa660-b80d-4dd2-a150-fee9352d72b8","Type":"ContainerDied","Data":"c8689fe53b6d835cb7d594850c99cc970b893f98dc80bad327aa74a59197c54d"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.751750 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8689fe53b6d835cb7d594850c99cc970b893f98dc80bad327aa74a59197c54d" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.751785 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e015-account-create-update-m75kd" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.763183 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"1e4f0e7898cde7a29868d7677f91a6fae395c27e58d4476c4df649466e26b91b"} Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.765557 4744 scope.go:117] "RemoveContainer" containerID="332907dabf85e7f3aafa53b6e966f55977756863ce4fc655555d149f5f4dfd8e" Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.815657 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fbvs9"] Dec 01 08:34:56 crc kubenswrapper[4744]: I1201 08:34:56.822090 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-fbvs9"] Dec 01 08:34:57 crc kubenswrapper[4744]: I1201 08:34:57.776348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jx6b8" event={"ID":"08cb1243-7645-41a5-b5d2-ea22230908c7","Type":"ContainerStarted","Data":"f601837c12e23fdf5cb5eeecf24667a4e5c865a0178787fd7f6e26a81fa80f09"} Dec 01 08:34:57 crc kubenswrapper[4744]: I1201 08:34:57.796557 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jx6b8" podStartSLOduration=3.075555152 podStartE2EDuration="23.796535841s" podCreationTimestamp="2025-12-01 08:34:34 +0000 UTC" firstStartedPulling="2025-12-01 08:34:35.440840266 +0000 UTC m=+1007.429898177" lastFinishedPulling="2025-12-01 08:34:56.161820945 +0000 UTC m=+1028.150878866" observedRunningTime="2025-12-01 08:34:57.790269231 +0000 UTC m=+1029.779327172" watchObservedRunningTime="2025-12-01 08:34:57.796535841 +0000 UTC m=+1029.785593762" Dec 01 08:34:58 crc kubenswrapper[4744]: I1201 08:34:58.301008 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3154b059-5545-416d-aca9-426f493b85f2" path="/var/lib/kubelet/pods/3154b059-5545-416d-aca9-426f493b85f2/volumes" Dec 01 08:35:00 crc kubenswrapper[4744]: I1201 08:35:00.808062 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d42bc" event={"ID":"d725b87c-b27b-414d-824d-e9f9f3172d69","Type":"ContainerStarted","Data":"03c20540d1d86beb58aaacb11728a294a18816f02865bcb699d8d6184cf66c6b"} Dec 01 08:35:00 crc kubenswrapper[4744]: I1201 08:35:00.833756 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-d42bc" podStartSLOduration=3.424821287 podStartE2EDuration="15.833735052s" podCreationTimestamp="2025-12-01 08:34:45 +0000 UTC" firstStartedPulling="2025-12-01 08:34:47.474629239 +0000 UTC m=+1019.463687160" lastFinishedPulling="2025-12-01 08:34:59.883543004 +0000 UTC m=+1031.872600925" observedRunningTime="2025-12-01 08:35:00.830401076 +0000 UTC m=+1032.819459037" watchObservedRunningTime="2025-12-01 08:35:00.833735052 +0000 UTC m=+1032.822792973" Dec 01 08:35:02 crc kubenswrapper[4744]: I1201 08:35:02.829129 4744 generic.go:334] "Generic (PLEG): container finished" podID="d725b87c-b27b-414d-824d-e9f9f3172d69" containerID="03c20540d1d86beb58aaacb11728a294a18816f02865bcb699d8d6184cf66c6b" exitCode=0 Dec 01 08:35:02 crc kubenswrapper[4744]: I1201 08:35:02.829271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d42bc" event={"ID":"d725b87c-b27b-414d-824d-e9f9f3172d69","Type":"ContainerDied","Data":"03c20540d1d86beb58aaacb11728a294a18816f02865bcb699d8d6184cf66c6b"} Dec 01 08:35:02 crc kubenswrapper[4744]: I1201 08:35:02.833358 4744 generic.go:334] "Generic (PLEG): container finished" podID="08cb1243-7645-41a5-b5d2-ea22230908c7" containerID="f601837c12e23fdf5cb5eeecf24667a4e5c865a0178787fd7f6e26a81fa80f09" exitCode=0 Dec 01 08:35:02 crc kubenswrapper[4744]: I1201 08:35:02.833472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jx6b8" event={"ID":"08cb1243-7645-41a5-b5d2-ea22230908c7","Type":"ContainerDied","Data":"f601837c12e23fdf5cb5eeecf24667a4e5c865a0178787fd7f6e26a81fa80f09"} Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.265320 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d42bc" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.358261 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jx6b8" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.408100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-combined-ca-bundle\") pod \"d725b87c-b27b-414d-824d-e9f9f3172d69\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.408549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggbpl\" (UniqueName: \"kubernetes.io/projected/d725b87c-b27b-414d-824d-e9f9f3172d69-kube-api-access-ggbpl\") pod \"d725b87c-b27b-414d-824d-e9f9f3172d69\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.408718 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-config-data\") pod \"d725b87c-b27b-414d-824d-e9f9f3172d69\" (UID: \"d725b87c-b27b-414d-824d-e9f9f3172d69\") " Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.415001 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d725b87c-b27b-414d-824d-e9f9f3172d69-kube-api-access-ggbpl" (OuterVolumeSpecName: "kube-api-access-ggbpl") pod "d725b87c-b27b-414d-824d-e9f9f3172d69" (UID: "d725b87c-b27b-414d-824d-e9f9f3172d69"). InnerVolumeSpecName "kube-api-access-ggbpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.439816 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d725b87c-b27b-414d-824d-e9f9f3172d69" (UID: "d725b87c-b27b-414d-824d-e9f9f3172d69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.460252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-config-data" (OuterVolumeSpecName: "config-data") pod "d725b87c-b27b-414d-824d-e9f9f3172d69" (UID: "d725b87c-b27b-414d-824d-e9f9f3172d69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.510509 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-db-sync-config-data\") pod \"08cb1243-7645-41a5-b5d2-ea22230908c7\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.510574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-config-data\") pod \"08cb1243-7645-41a5-b5d2-ea22230908c7\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.510643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-combined-ca-bundle\") pod \"08cb1243-7645-41a5-b5d2-ea22230908c7\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.510699 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9kzj\" (UniqueName: \"kubernetes.io/projected/08cb1243-7645-41a5-b5d2-ea22230908c7-kube-api-access-m9kzj\") pod \"08cb1243-7645-41a5-b5d2-ea22230908c7\" (UID: \"08cb1243-7645-41a5-b5d2-ea22230908c7\") " Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.511138 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.511172 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d725b87c-b27b-414d-824d-e9f9f3172d69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.511193 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggbpl\" (UniqueName: \"kubernetes.io/projected/d725b87c-b27b-414d-824d-e9f9f3172d69-kube-api-access-ggbpl\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.513736 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08cb1243-7645-41a5-b5d2-ea22230908c7-kube-api-access-m9kzj" (OuterVolumeSpecName: "kube-api-access-m9kzj") pod "08cb1243-7645-41a5-b5d2-ea22230908c7" (UID: "08cb1243-7645-41a5-b5d2-ea22230908c7"). InnerVolumeSpecName "kube-api-access-m9kzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.514095 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "08cb1243-7645-41a5-b5d2-ea22230908c7" (UID: "08cb1243-7645-41a5-b5d2-ea22230908c7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.543812 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08cb1243-7645-41a5-b5d2-ea22230908c7" (UID: "08cb1243-7645-41a5-b5d2-ea22230908c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.597667 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-config-data" (OuterVolumeSpecName: "config-data") pod "08cb1243-7645-41a5-b5d2-ea22230908c7" (UID: "08cb1243-7645-41a5-b5d2-ea22230908c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.612703 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.612735 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.612745 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cb1243-7645-41a5-b5d2-ea22230908c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.612756 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9kzj\" (UniqueName: \"kubernetes.io/projected/08cb1243-7645-41a5-b5d2-ea22230908c7-kube-api-access-m9kzj\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.863972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d42bc" event={"ID":"d725b87c-b27b-414d-824d-e9f9f3172d69","Type":"ContainerDied","Data":"c633121d94932e5597e768674e81acccaa39c466112e570daa965193ec53a20d"} Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.864008 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c633121d94932e5597e768674e81acccaa39c466112e570daa965193ec53a20d" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.864290 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d42bc" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.866016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jx6b8" event={"ID":"08cb1243-7645-41a5-b5d2-ea22230908c7","Type":"ContainerDied","Data":"c3b8e2dae6b6ddc8b09fdc7a0a0f79261e52d199216bb3b37be4df0a211501c0"} Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.866036 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3b8e2dae6b6ddc8b09fdc7a0a0f79261e52d199216bb3b37be4df0a211501c0" Dec 01 08:35:04 crc kubenswrapper[4744]: I1201 08:35:04.866103 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jx6b8" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.195689 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nhv7l"] Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196053 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb2f8c0-0676-4f02-af2c-b4000c09ab26" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196069 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb2f8c0-0676-4f02-af2c-b4000c09ab26" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196081 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5076354c-7f79-490f-9ea7-c5d2dd60c545" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196088 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5076354c-7f79-490f-9ea7-c5d2dd60c545" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196100 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d725b87c-b27b-414d-824d-e9f9f3172d69" containerName="keystone-db-sync" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196106 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d725b87c-b27b-414d-824d-e9f9f3172d69" containerName="keystone-db-sync" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196120 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3154b059-5545-416d-aca9-426f493b85f2" containerName="init" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196125 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3154b059-5545-416d-aca9-426f493b85f2" containerName="init" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196133 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08cb1243-7645-41a5-b5d2-ea22230908c7" containerName="glance-db-sync" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196138 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="08cb1243-7645-41a5-b5d2-ea22230908c7" containerName="glance-db-sync" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196145 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b116a0ec-0413-4532-b820-181cbf9f3515" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196151 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b116a0ec-0413-4532-b820-181cbf9f3515" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196164 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac6bc34-6d3b-4860-aab8-bdf00a485793" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196169 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac6bc34-6d3b-4860-aab8-bdf00a485793" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196179 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4341bae7-8d88-4d04-8670-43125be4a7a3" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196186 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4341bae7-8d88-4d04-8670-43125be4a7a3" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196198 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198a3f7b-ed68-4dee-92e1-9aa0e9362304" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196204 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="198a3f7b-ed68-4dee-92e1-9aa0e9362304" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196215 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9aa660-b80d-4dd2-a150-fee9352d72b8" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196220 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9aa660-b80d-4dd2-a150-fee9352d72b8" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196230 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e4d4551-0e28-432e-bd4e-e783133b0a4b" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196236 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e4d4551-0e28-432e-bd4e-e783133b0a4b" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.196248 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3154b059-5545-416d-aca9-426f493b85f2" containerName="dnsmasq-dns" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196253 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3154b059-5545-416d-aca9-426f493b85f2" containerName="dnsmasq-dns" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196393 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3154b059-5545-416d-aca9-426f493b85f2" containerName="dnsmasq-dns" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196418 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4341bae7-8d88-4d04-8670-43125be4a7a3" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196429 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d725b87c-b27b-414d-824d-e9f9f3172d69" containerName="keystone-db-sync" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196439 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ac6bc34-6d3b-4860-aab8-bdf00a485793" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196452 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9aa660-b80d-4dd2-a150-fee9352d72b8" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196473 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5076354c-7f79-490f-9ea7-c5d2dd60c545" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196481 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="198a3f7b-ed68-4dee-92e1-9aa0e9362304" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196491 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e4d4551-0e28-432e-bd4e-e783133b0a4b" containerName="mariadb-account-create-update" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196503 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="08cb1243-7645-41a5-b5d2-ea22230908c7" containerName="glance-db-sync" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196515 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fb2f8c0-0676-4f02-af2c-b4000c09ab26" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.196525 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b116a0ec-0413-4532-b820-181cbf9f3515" containerName="mariadb-database-create" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.197099 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.200656 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4jgn6" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.202244 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.202301 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.202329 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.202385 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.204800 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-9dzb2"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.206311 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.320512 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nhv7l"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-combined-ca-bundle\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348638 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmfwx\" (UniqueName: \"kubernetes.io/projected/bce6df6c-65f0-4984-8887-4378ec0b59ea-kube-api-access-kmfwx\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-credential-keys\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348786 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-config-data\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-scripts\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348864 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qgc8\" (UniqueName: \"kubernetes.io/projected/c6556f94-84b8-4cbc-b93f-464df38b5a15-kube-api-access-2qgc8\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-fernet-keys\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348952 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-config\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.348990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.364060 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-9dzb2"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.409335 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-x664l"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.410320 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.420694 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.421956 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-qjzlh" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451318 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451378 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-credential-keys\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451440 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-config-data\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451464 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-scripts\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451524 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qgc8\" (UniqueName: \"kubernetes.io/projected/c6556f94-84b8-4cbc-b93f-464df38b5a15-kube-api-access-2qgc8\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451550 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451572 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-fernet-keys\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-config\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451678 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451738 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-combined-ca-bundle\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.451764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmfwx\" (UniqueName: \"kubernetes.io/projected/bce6df6c-65f0-4984-8887-4378ec0b59ea-kube-api-access-kmfwx\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.455653 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.456899 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.457916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-config\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.458562 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.462271 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.465473 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-x664l"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.480143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-config-data\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.480781 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmfwx\" (UniqueName: \"kubernetes.io/projected/bce6df6c-65f0-4984-8887-4378ec0b59ea-kube-api-access-kmfwx\") pod \"dnsmasq-dns-6f8c45789f-9dzb2\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.484797 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-fernet-keys\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.485133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-credential-keys\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.491100 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-combined-ca-bundle\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.504630 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qgc8\" (UniqueName: \"kubernetes.io/projected/c6556f94-84b8-4cbc-b93f-464df38b5a15-kube-api-access-2qgc8\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.527253 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-scripts\") pod \"keystone-bootstrap-nhv7l\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.560816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nj56\" (UniqueName: \"kubernetes.io/projected/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-kube-api-access-4nj56\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.560929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-combined-ca-bundle\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.560975 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-config-data\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.582327 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-5k8xk"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.583396 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.594619 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.595014 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.595210 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4wwpn" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.599765 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-t2mxc"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.601594 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.618946 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.619473 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-sln7j" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.619799 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.620311 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5k8xk"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.622790 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.623437 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.624671 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-t2mxc"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.663221 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-combined-ca-bundle\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.663560 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-config-data\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.663616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nj56\" (UniqueName: \"kubernetes.io/projected/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-kube-api-access-4nj56\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.670577 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-combined-ca-bundle\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.685143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-config-data\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.698716 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nj56\" (UniqueName: \"kubernetes.io/projected/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-kube-api-access-4nj56\") pod \"heat-db-sync-x664l\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.703598 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-bzxk9"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.704758 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.710897 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-bzxk9"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.712221 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.712528 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.722143 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4dm2j" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.727344 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.729434 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.731129 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.737570 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.756991 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-9dzb2"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.759386 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-x664l" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765220 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-scripts\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-combined-ca-bundle\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765298 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-config\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-combined-ca-bundle\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765360 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq2xc\" (UniqueName: \"kubernetes.io/projected/d51f040d-3a61-47f1-8699-c88fbaabdee5-kube-api-access-jq2xc\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765394 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3718521-a22d-4350-aa32-98987bfe5632-etc-machine-id\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765430 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-db-sync-config-data\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765455 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv46w\" (UniqueName: \"kubernetes.io/projected/d3718521-a22d-4350-aa32-98987bfe5632-kube-api-access-wv46w\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.765471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-config-data\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.767520 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.819670 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-4bhng"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.821055 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.837478 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-4bhng"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.872877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-db-sync-config-data\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.872917 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-scripts\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.872953 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-combined-ca-bundle\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.872973 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-config-data\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.872995 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8psvl\" (UniqueName: \"kubernetes.io/projected/95085c33-3a8b-4af5-aaef-bd718dabcf6f-kube-api-access-8psvl\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-scripts\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873026 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv46w\" (UniqueName: \"kubernetes.io/projected/d3718521-a22d-4350-aa32-98987bfe5632-kube-api-access-wv46w\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873044 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-config-data\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-scripts\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873108 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-combined-ca-bundle\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-config\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873164 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95085c33-3a8b-4af5-aaef-bd718dabcf6f-logs\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873188 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-combined-ca-bundle\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgrq9\" (UniqueName: \"kubernetes.io/projected/96a23a18-05c4-409a-950d-0743db8fc652-kube-api-access-wgrq9\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873225 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-run-httpd\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-config-data\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq2xc\" (UniqueName: \"kubernetes.io/projected/d51f040d-3a61-47f1-8699-c88fbaabdee5-kube-api-access-jq2xc\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873279 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-log-httpd\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873313 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3718521-a22d-4350-aa32-98987bfe5632-etc-machine-id\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.873337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.886107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-combined-ca-bundle\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.886489 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3718521-a22d-4350-aa32-98987bfe5632-etc-machine-id\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.893624 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-8mwhk"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.893920 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-db-sync-config-data\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.894390 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-scripts\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.894694 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.900332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-config\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.903698 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.903914 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mgtfz" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.912170 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-config-data\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.922820 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-combined-ca-bundle\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.927116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv46w\" (UniqueName: \"kubernetes.io/projected/d3718521-a22d-4350-aa32-98987bfe5632-kube-api-access-wv46w\") pod \"cinder-db-sync-t2mxc\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.940653 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-4bhng"] Dec 01 08:35:05 crc kubenswrapper[4744]: E1201 08:35:05.941229 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-8v6tn ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" podUID="8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.941668 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.941792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq2xc\" (UniqueName: \"kubernetes.io/projected/d51f040d-3a61-47f1-8699-c88fbaabdee5-kube-api-access-jq2xc\") pod \"neutron-db-sync-5k8xk\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.960980 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8mwhk"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974498 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgrq9\" (UniqueName: \"kubernetes.io/projected/96a23a18-05c4-409a-950d-0743db8fc652-kube-api-access-wgrq9\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-run-httpd\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974602 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-config-data\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974619 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-log-httpd\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974643 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-config\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974705 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974737 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9jth\" (UniqueName: \"kubernetes.io/projected/ed2a24ae-749f-4faf-b077-060df14120bb-kube-api-access-m9jth\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-scripts\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-combined-ca-bundle\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-config-data\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8psvl\" (UniqueName: \"kubernetes.io/projected/95085c33-3a8b-4af5-aaef-bd718dabcf6f-kube-api-access-8psvl\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-scripts\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974899 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v6tn\" (UniqueName: \"kubernetes.io/projected/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-kube-api-access-8v6tn\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-combined-ca-bundle\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.974987 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-run-httpd\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.975063 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.975099 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95085c33-3a8b-4af5-aaef-bd718dabcf6f-logs\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.975124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-db-sync-config-data\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.983829 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-log-httpd\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.986200 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95085c33-3a8b-4af5-aaef-bd718dabcf6f-logs\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.989520 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-5d7w6"] Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.995817 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:05 crc kubenswrapper[4744]: I1201 08:35:05.998028 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-5d7w6"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.015342 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-config-data\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.026606 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-scripts\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.027123 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-combined-ca-bundle\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.027908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-config-data\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.044531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.049519 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgrq9\" (UniqueName: \"kubernetes.io/projected/96a23a18-05c4-409a-950d-0743db8fc652-kube-api-access-wgrq9\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.049983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-scripts\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.050162 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " pod="openstack/ceilometer-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.055849 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8psvl\" (UniqueName: \"kubernetes.io/projected/95085c33-3a8b-4af5-aaef-bd718dabcf6f-kube-api-access-8psvl\") pod \"placement-db-sync-bzxk9\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.079336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.079376 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-config\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.079397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.079435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-db-sync-config-data\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.079457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8lhc\" (UniqueName: \"kubernetes.io/projected/9969b75a-1def-4af6-a70b-a89129668ffa-kube-api-access-t8lhc\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.079488 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.079508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-config\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080627 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080724 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9jth\" (UniqueName: \"kubernetes.io/projected/ed2a24ae-749f-4faf-b077-060df14120bb-kube-api-access-m9jth\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080782 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v6tn\" (UniqueName: \"kubernetes.io/projected/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-kube-api-access-8v6tn\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-combined-ca-bundle\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.080882 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.082111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.083648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-config\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.084538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.084787 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.086921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.093179 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-db-sync-config-data\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.094023 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-combined-ca-bundle\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.113031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9jth\" (UniqueName: \"kubernetes.io/projected/ed2a24ae-749f-4faf-b077-060df14120bb-kube-api-access-m9jth\") pod \"barbican-db-sync-8mwhk\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.113339 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.130314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v6tn\" (UniqueName: \"kubernetes.io/projected/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-kube-api-access-8v6tn\") pod \"dnsmasq-dns-fcfdd6f9f-4bhng\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.156828 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.182437 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-config\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.182716 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.182750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8lhc\" (UniqueName: \"kubernetes.io/projected/9969b75a-1def-4af6-a70b-a89129668ffa-kube-api-access-t8lhc\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.182786 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.182809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.182869 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.183612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.184091 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-config\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.184599 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.185302 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.187110 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.206862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.210175 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8lhc\" (UniqueName: \"kubernetes.io/projected/9969b75a-1def-4af6-a70b-a89129668ffa-kube-api-access-t8lhc\") pod \"dnsmasq-dns-57c957c4ff-5d7w6\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.233176 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.403164 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.436255 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nhv7l"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.447769 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-9dzb2"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.585448 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.586813 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.596835 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.597448 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.597663 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gl4sq" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.605355 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.673779 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-t2mxc"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.692315 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-logs\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.692375 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.692443 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.692551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.692594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-scripts\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.692637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-config-data\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.692664 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh77h\" (UniqueName: \"kubernetes.io/projected/9efef428-22bb-47d2-a871-18cdc5495886-kube-api-access-lh77h\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.707711 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-x664l"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.794593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.794669 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-scripts\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.794722 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-config-data\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.794750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh77h\" (UniqueName: \"kubernetes.io/projected/9efef428-22bb-47d2-a871-18cdc5495886-kube-api-access-lh77h\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.794787 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-logs\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.794816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.794848 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.797931 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.800598 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.800635 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-logs\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.802175 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.809020 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-scripts\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.817773 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-config-data\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.827131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh77h\" (UniqueName: \"kubernetes.io/projected/9efef428-22bb-47d2-a871-18cdc5495886-kube-api-access-lh77h\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.843993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.866021 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.926910 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.975335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:06 crc kubenswrapper[4744]: I1201 08:35:06.985460 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5k8xk"] Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.002367 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.013546 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-bzxk9"] Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.036471 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.037969 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.040810 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.051322 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.059233 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8mwhk"] Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.105851 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-config\") pod \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.105908 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-sb\") pod \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.105938 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-nb\") pod \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106008 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-swift-storage-0\") pod \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106082 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-svc\") pod \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106107 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v6tn\" (UniqueName: \"kubernetes.io/projected/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-kube-api-access-8v6tn\") pod \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\" (UID: \"8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63\") " Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106338 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106364 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-logs\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106429 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106447 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106490 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49hds\" (UniqueName: \"kubernetes.io/projected/22235e5d-c2cb-473f-a7aa-06f4ff712608-kube-api-access-49hds\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.106963 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-config" (OuterVolumeSpecName: "config") pod "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" (UID: "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.107461 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" (UID: "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.107886 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" (UID: "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.108100 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" (UID: "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.110568 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" (UID: "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.112241 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-kube-api-access-8v6tn" (OuterVolumeSpecName: "kube-api-access-8v6tn") pod "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" (UID: "8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63"). InnerVolumeSpecName "kube-api-access-8v6tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.208778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.208871 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49hds\" (UniqueName: \"kubernetes.io/projected/22235e5d-c2cb-473f-a7aa-06f4ff712608-kube-api-access-49hds\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-logs\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209140 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209242 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209282 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209295 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v6tn\" (UniqueName: \"kubernetes.io/projected/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-kube-api-access-8v6tn\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209309 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209319 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.209346 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.210105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.210144 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-logs\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.210535 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.213599 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.215256 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.216187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.229308 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49hds\" (UniqueName: \"kubernetes.io/projected/22235e5d-c2cb-473f-a7aa-06f4ff712608-kube-api-access-49hds\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.252072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.375623 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.506140 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.553948 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.603143 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:07 crc kubenswrapper[4744]: W1201 08:35:07.822400 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbce6df6c_65f0_4984_8887_4378ec0b59ea.slice/crio-21cdd5226b547d9c6d9a5b1d59dd7916f46869adc422d827b5d260107eea8ea1 WatchSource:0}: Error finding container 21cdd5226b547d9c6d9a5b1d59dd7916f46869adc422d827b5d260107eea8ea1: Status 404 returned error can't find the container with id 21cdd5226b547d9c6d9a5b1d59dd7916f46869adc422d827b5d260107eea8ea1 Dec 01 08:35:07 crc kubenswrapper[4744]: W1201 08:35:07.836239 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd51f040d_3a61_47f1_8699_c88fbaabdee5.slice/crio-48a3d190ed79a2c929bbf02730fa7b78396a4cdc7c011028b689198155f76f1c WatchSource:0}: Error finding container 48a3d190ed79a2c929bbf02730fa7b78396a4cdc7c011028b689198155f76f1c: Status 404 returned error can't find the container with id 48a3d190ed79a2c929bbf02730fa7b78396a4cdc7c011028b689198155f76f1c Dec 01 08:35:07 crc kubenswrapper[4744]: I1201 08:35:07.995088 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-x664l" event={"ID":"ab3b54ed-348a-4314-8e92-3e977ee8f1ff","Type":"ContainerStarted","Data":"defe48eadc0573a912670898c2992ef4c6762f0c03f98124303c4624e374650c"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.000316 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bzxk9" event={"ID":"95085c33-3a8b-4af5-aaef-bd718dabcf6f","Type":"ContainerStarted","Data":"56a818bac339ebe28970e1dc9ae87d8307a49bebc8a7f595d78f7b1653a97d4c"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.013435 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhv7l" event={"ID":"c6556f94-84b8-4cbc-b93f-464df38b5a15","Type":"ContainerStarted","Data":"c37179e9d623fbdf9d2a1c3e5c8c4cccf1178b476430539fafa1694c48e24820"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.017933 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2mxc" event={"ID":"d3718521-a22d-4350-aa32-98987bfe5632","Type":"ContainerStarted","Data":"79de33a4f6fc4753edcae6ad455721d3e9d5338c1321747dc34489f7bac58534"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.020748 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerStarted","Data":"669a774253f2ff55d515414d17783a1443beff590a5ab59ad4798a76bbec3fc6"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.030948 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8mwhk" event={"ID":"ed2a24ae-749f-4faf-b077-060df14120bb","Type":"ContainerStarted","Data":"f308ec7a37d15fbcb510a5235947b4a8530ab5f7ec41c1a6ee39233bcaea93d2"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.054317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5k8xk" event={"ID":"d51f040d-3a61-47f1-8699-c88fbaabdee5","Type":"ContainerStarted","Data":"48a3d190ed79a2c929bbf02730fa7b78396a4cdc7c011028b689198155f76f1c"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.065776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" event={"ID":"bce6df6c-65f0-4984-8887-4378ec0b59ea","Type":"ContainerStarted","Data":"21cdd5226b547d9c6d9a5b1d59dd7916f46869adc422d827b5d260107eea8ea1"} Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.065817 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-4bhng" Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.142302 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-4bhng"] Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.148040 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-4bhng"] Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.298477 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63" path="/var/lib/kubelet/pods/8c1bf1fb-dbe1-446c-af7c-6f11bc4d2c63/volumes" Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.517696 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-5d7w6"] Dec 01 08:35:08 crc kubenswrapper[4744]: I1201 08:35:08.608167 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:09 crc kubenswrapper[4744]: I1201 08:35:09.074501 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" event={"ID":"9969b75a-1def-4af6-a70b-a89129668ffa","Type":"ContainerStarted","Data":"8d8d0e6247b1dacd9c56c9be9abdba58b8b423643a2a934a534a270194950a5b"} Dec 01 08:35:09 crc kubenswrapper[4744]: I1201 08:35:09.082009 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9efef428-22bb-47d2-a871-18cdc5495886","Type":"ContainerStarted","Data":"f6d81974aeac79d9c11f5e9f4cf2b4dfe184e58bcb2d3c35498ce29e6bd0337e"} Dec 01 08:35:09 crc kubenswrapper[4744]: I1201 08:35:09.444118 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:09 crc kubenswrapper[4744]: W1201 08:35:09.450754 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22235e5d_c2cb_473f_a7aa_06f4ff712608.slice/crio-985a083bd3f10cf2b97cecea7493e621d55e45ce81330f3126de11d34e9d84dd WatchSource:0}: Error finding container 985a083bd3f10cf2b97cecea7493e621d55e45ce81330f3126de11d34e9d84dd: Status 404 returned error can't find the container with id 985a083bd3f10cf2b97cecea7493e621d55e45ce81330f3126de11d34e9d84dd Dec 01 08:35:10 crc kubenswrapper[4744]: I1201 08:35:10.093678 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22235e5d-c2cb-473f-a7aa-06f4ff712608","Type":"ContainerStarted","Data":"985a083bd3f10cf2b97cecea7493e621d55e45ce81330f3126de11d34e9d84dd"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.134169 4744 generic.go:334] "Generic (PLEG): container finished" podID="bce6df6c-65f0-4984-8887-4378ec0b59ea" containerID="e287d471096b4ae9c5299db1506ab1bdfacd16b7ca682823c8a4d728368df5e2" exitCode=0 Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.134670 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" event={"ID":"bce6df6c-65f0-4984-8887-4378ec0b59ea","Type":"ContainerDied","Data":"e287d471096b4ae9c5299db1506ab1bdfacd16b7ca682823c8a4d728368df5e2"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.164004 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9efef428-22bb-47d2-a871-18cdc5495886","Type":"ContainerStarted","Data":"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.164476 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9efef428-22bb-47d2-a871-18cdc5495886","Type":"ContainerStarted","Data":"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.164603 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-log" containerID="cri-o://175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933" gracePeriod=30 Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.164699 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-httpd" containerID="cri-o://a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b" gracePeriod=30 Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.179774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22235e5d-c2cb-473f-a7aa-06f4ff712608","Type":"ContainerStarted","Data":"60c9039618c0dce314616a46e3c1aac7ee9e8543ccc804a91bf68e1160939c6f"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.179821 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22235e5d-c2cb-473f-a7aa-06f4ff712608","Type":"ContainerStarted","Data":"ae6589dc2f4b0ddedb8a62ece74df90a775d331578f6c67d02ca20c3228d0db3"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.179962 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-log" containerID="cri-o://ae6589dc2f4b0ddedb8a62ece74df90a775d331578f6c67d02ca20c3228d0db3" gracePeriod=30 Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.180278 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-httpd" containerID="cri-o://60c9039618c0dce314616a46e3c1aac7ee9e8543ccc804a91bf68e1160939c6f" gracePeriod=30 Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.199164 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhv7l" event={"ID":"c6556f94-84b8-4cbc-b93f-464df38b5a15","Type":"ContainerStarted","Data":"b92d0cd7d77d5e49e122180332b578460f2a07dc39bfee62e14671c8a66b2d78"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.222149 4744 generic.go:334] "Generic (PLEG): container finished" podID="9969b75a-1def-4af6-a70b-a89129668ffa" containerID="6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5" exitCode=0 Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.222223 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" event={"ID":"9969b75a-1def-4af6-a70b-a89129668ffa","Type":"ContainerDied","Data":"6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.223421 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.223370801 podStartE2EDuration="6.223370801s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:11.211483562 +0000 UTC m=+1043.200541483" watchObservedRunningTime="2025-12-01 08:35:11.223370801 +0000 UTC m=+1043.212428712" Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.242631 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5k8xk" event={"ID":"d51f040d-3a61-47f1-8699-c88fbaabdee5","Type":"ContainerStarted","Data":"3c4190b993a15f411262d140d6dc55defae79c915dc8bbfafcb13a406c81c171"} Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.257920 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.257899434 podStartE2EDuration="6.257899434s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:11.235053933 +0000 UTC m=+1043.224111854" watchObservedRunningTime="2025-12-01 08:35:11.257899434 +0000 UTC m=+1043.246957355" Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.337455 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nhv7l" podStartSLOduration=6.337439258 podStartE2EDuration="6.337439258s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:11.312706874 +0000 UTC m=+1043.301764795" watchObservedRunningTime="2025-12-01 08:35:11.337439258 +0000 UTC m=+1043.326497179" Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.356755 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-5k8xk" podStartSLOduration=6.356735878 podStartE2EDuration="6.356735878s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:11.338697484 +0000 UTC m=+1043.327755405" watchObservedRunningTime="2025-12-01 08:35:11.356735878 +0000 UTC m=+1043.345793799" Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.715506 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.834993 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-nb\") pod \"bce6df6c-65f0-4984-8887-4378ec0b59ea\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.835344 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-swift-storage-0\") pod \"bce6df6c-65f0-4984-8887-4378ec0b59ea\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.835495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-svc\") pod \"bce6df6c-65f0-4984-8887-4378ec0b59ea\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.835551 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-config\") pod \"bce6df6c-65f0-4984-8887-4378ec0b59ea\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.835600 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmfwx\" (UniqueName: \"kubernetes.io/projected/bce6df6c-65f0-4984-8887-4378ec0b59ea-kube-api-access-kmfwx\") pod \"bce6df6c-65f0-4984-8887-4378ec0b59ea\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.835628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-sb\") pod \"bce6df6c-65f0-4984-8887-4378ec0b59ea\" (UID: \"bce6df6c-65f0-4984-8887-4378ec0b59ea\") " Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.866830 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bce6df6c-65f0-4984-8887-4378ec0b59ea-kube-api-access-kmfwx" (OuterVolumeSpecName: "kube-api-access-kmfwx") pod "bce6df6c-65f0-4984-8887-4378ec0b59ea" (UID: "bce6df6c-65f0-4984-8887-4378ec0b59ea"). InnerVolumeSpecName "kube-api-access-kmfwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:11 crc kubenswrapper[4744]: I1201 08:35:11.869809 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bce6df6c-65f0-4984-8887-4378ec0b59ea" (UID: "bce6df6c-65f0-4984-8887-4378ec0b59ea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.907908 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bce6df6c-65f0-4984-8887-4378ec0b59ea" (UID: "bce6df6c-65f0-4984-8887-4378ec0b59ea"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.908064 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-config" (OuterVolumeSpecName: "config") pod "bce6df6c-65f0-4984-8887-4378ec0b59ea" (UID: "bce6df6c-65f0-4984-8887-4378ec0b59ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.908712 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bce6df6c-65f0-4984-8887-4378ec0b59ea" (UID: "bce6df6c-65f0-4984-8887-4378ec0b59ea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.909743 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bce6df6c-65f0-4984-8887-4378ec0b59ea" (UID: "bce6df6c-65f0-4984-8887-4378ec0b59ea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.938734 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.938762 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.938773 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmfwx\" (UniqueName: \"kubernetes.io/projected/bce6df6c-65f0-4984-8887-4378ec0b59ea-kube-api-access-kmfwx\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.938783 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.938791 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:11.938799 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bce6df6c-65f0-4984-8887-4378ec0b59ea-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.100247 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.147949 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-logs\") pod \"9efef428-22bb-47d2-a871-18cdc5495886\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.148035 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"9efef428-22bb-47d2-a871-18cdc5495886\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.148133 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-config-data\") pod \"9efef428-22bb-47d2-a871-18cdc5495886\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.148170 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-combined-ca-bundle\") pod \"9efef428-22bb-47d2-a871-18cdc5495886\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.148216 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-httpd-run\") pod \"9efef428-22bb-47d2-a871-18cdc5495886\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.148247 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh77h\" (UniqueName: \"kubernetes.io/projected/9efef428-22bb-47d2-a871-18cdc5495886-kube-api-access-lh77h\") pod \"9efef428-22bb-47d2-a871-18cdc5495886\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.148275 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-scripts\") pod \"9efef428-22bb-47d2-a871-18cdc5495886\" (UID: \"9efef428-22bb-47d2-a871-18cdc5495886\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.150846 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9efef428-22bb-47d2-a871-18cdc5495886" (UID: "9efef428-22bb-47d2-a871-18cdc5495886"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.156544 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-scripts" (OuterVolumeSpecName: "scripts") pod "9efef428-22bb-47d2-a871-18cdc5495886" (UID: "9efef428-22bb-47d2-a871-18cdc5495886"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.167539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-logs" (OuterVolumeSpecName: "logs") pod "9efef428-22bb-47d2-a871-18cdc5495886" (UID: "9efef428-22bb-47d2-a871-18cdc5495886"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.170285 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "9efef428-22bb-47d2-a871-18cdc5495886" (UID: "9efef428-22bb-47d2-a871-18cdc5495886"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.173394 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efef428-22bb-47d2-a871-18cdc5495886-kube-api-access-lh77h" (OuterVolumeSpecName: "kube-api-access-lh77h") pod "9efef428-22bb-47d2-a871-18cdc5495886" (UID: "9efef428-22bb-47d2-a871-18cdc5495886"). InnerVolumeSpecName "kube-api-access-lh77h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.199647 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9efef428-22bb-47d2-a871-18cdc5495886" (UID: "9efef428-22bb-47d2-a871-18cdc5495886"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.244976 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-config-data" (OuterVolumeSpecName: "config-data") pod "9efef428-22bb-47d2-a871-18cdc5495886" (UID: "9efef428-22bb-47d2-a871-18cdc5495886"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.250175 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.250508 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.250519 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.250531 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh77h\" (UniqueName: \"kubernetes.io/projected/9efef428-22bb-47d2-a871-18cdc5495886-kube-api-access-lh77h\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.250541 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efef428-22bb-47d2-a871-18cdc5495886-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.250552 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efef428-22bb-47d2-a871-18cdc5495886-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.250579 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.261455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" event={"ID":"9969b75a-1def-4af6-a70b-a89129668ffa","Type":"ContainerStarted","Data":"74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228"} Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.261729 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.267030 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" event={"ID":"bce6df6c-65f0-4984-8887-4378ec0b59ea","Type":"ContainerDied","Data":"21cdd5226b547d9c6d9a5b1d59dd7916f46869adc422d827b5d260107eea8ea1"} Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.267076 4744 scope.go:117] "RemoveContainer" containerID="e287d471096b4ae9c5299db1506ab1bdfacd16b7ca682823c8a4d728368df5e2" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.267216 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-9dzb2" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.275386 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.278330 4744 generic.go:334] "Generic (PLEG): container finished" podID="9efef428-22bb-47d2-a871-18cdc5495886" containerID="a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b" exitCode=0 Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.278357 4744 generic.go:334] "Generic (PLEG): container finished" podID="9efef428-22bb-47d2-a871-18cdc5495886" containerID="175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933" exitCode=143 Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.278542 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9efef428-22bb-47d2-a871-18cdc5495886","Type":"ContainerDied","Data":"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b"} Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.278584 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9efef428-22bb-47d2-a871-18cdc5495886","Type":"ContainerDied","Data":"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933"} Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.278601 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9efef428-22bb-47d2-a871-18cdc5495886","Type":"ContainerDied","Data":"f6d81974aeac79d9c11f5e9f4cf2b4dfe184e58bcb2d3c35498ce29e6bd0337e"} Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.278685 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.293664 4744 generic.go:334] "Generic (PLEG): container finished" podID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerID="60c9039618c0dce314616a46e3c1aac7ee9e8543ccc804a91bf68e1160939c6f" exitCode=0 Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.293694 4744 generic.go:334] "Generic (PLEG): container finished" podID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerID="ae6589dc2f4b0ddedb8a62ece74df90a775d331578f6c67d02ca20c3228d0db3" exitCode=143 Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.296908 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" podStartSLOduration=7.296879556 podStartE2EDuration="7.296879556s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:12.296733582 +0000 UTC m=+1044.285791503" watchObservedRunningTime="2025-12-01 08:35:12.296879556 +0000 UTC m=+1044.285937477" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.331018 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22235e5d-c2cb-473f-a7aa-06f4ff712608","Type":"ContainerDied","Data":"60c9039618c0dce314616a46e3c1aac7ee9e8543ccc804a91bf68e1160939c6f"} Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.331061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22235e5d-c2cb-473f-a7aa-06f4ff712608","Type":"ContainerDied","Data":"ae6589dc2f4b0ddedb8a62ece74df90a775d331578f6c67d02ca20c3228d0db3"} Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.353371 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.359650 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-9dzb2"] Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.368189 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-9dzb2"] Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.379566 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.381163 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.453351 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:12 crc kubenswrapper[4744]: E1201 08:35:12.454241 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-httpd" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.454256 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-httpd" Dec 01 08:35:12 crc kubenswrapper[4744]: E1201 08:35:12.454281 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-log" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.454288 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-log" Dec 01 08:35:12 crc kubenswrapper[4744]: E1201 08:35:12.454332 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bce6df6c-65f0-4984-8887-4378ec0b59ea" containerName="init" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.454339 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bce6df6c-65f0-4984-8887-4378ec0b59ea" containerName="init" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.454752 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-httpd" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.454779 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bce6df6c-65f0-4984-8887-4378ec0b59ea" containerName="init" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.454795 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efef428-22bb-47d2-a871-18cdc5495886" containerName="glance-log" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.458459 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.462300 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.480320 4744 scope.go:117] "RemoveContainer" containerID="a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.491561 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.533694 4744 scope.go:117] "RemoveContainer" containerID="175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.563953 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-config-data\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.564015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2tpq\" (UniqueName: \"kubernetes.io/projected/abf48605-5626-4522-bab4-63b0ff51943a-kube-api-access-h2tpq\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.564042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-scripts\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.564067 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.564121 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.564203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.564225 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-logs\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.589075 4744 scope.go:117] "RemoveContainer" containerID="a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b" Dec 01 08:35:12 crc kubenswrapper[4744]: E1201 08:35:12.592355 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b\": container with ID starting with a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b not found: ID does not exist" containerID="a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.592385 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b"} err="failed to get container status \"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b\": rpc error: code = NotFound desc = could not find container \"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b\": container with ID starting with a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b not found: ID does not exist" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.592469 4744 scope.go:117] "RemoveContainer" containerID="175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933" Dec 01 08:35:12 crc kubenswrapper[4744]: E1201 08:35:12.593200 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933\": container with ID starting with 175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933 not found: ID does not exist" containerID="175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.593246 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933"} err="failed to get container status \"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933\": rpc error: code = NotFound desc = could not find container \"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933\": container with ID starting with 175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933 not found: ID does not exist" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.593276 4744 scope.go:117] "RemoveContainer" containerID="a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.593707 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b"} err="failed to get container status \"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b\": rpc error: code = NotFound desc = could not find container \"a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b\": container with ID starting with a63604f04280b5e0ce7e7f27462bd84fe61b27a8b7fd089ca83f63b75213fa3b not found: ID does not exist" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.593723 4744 scope.go:117] "RemoveContainer" containerID="175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.596973 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933"} err="failed to get container status \"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933\": rpc error: code = NotFound desc = could not find container \"175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933\": container with ID starting with 175005e968d4a3f1cc91c90ad8a8f8e8c41d7c9a5ccd5160f3d5f7fa0daec933 not found: ID does not exist" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.665890 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-config-data\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.665945 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2tpq\" (UniqueName: \"kubernetes.io/projected/abf48605-5626-4522-bab4-63b0ff51943a-kube-api-access-h2tpq\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.665964 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-scripts\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.665982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.666021 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.666084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.666103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-logs\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.666440 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.666511 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-logs\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.671170 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.672111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-config-data\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.672626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.683389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-scripts\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.694087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2tpq\" (UniqueName: \"kubernetes.io/projected/abf48605-5626-4522-bab4-63b0ff51943a-kube-api-access-h2tpq\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.711921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.791840 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.944578 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.970868 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-config-data\") pod \"22235e5d-c2cb-473f-a7aa-06f4ff712608\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.970972 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49hds\" (UniqueName: \"kubernetes.io/projected/22235e5d-c2cb-473f-a7aa-06f4ff712608-kube-api-access-49hds\") pod \"22235e5d-c2cb-473f-a7aa-06f4ff712608\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.971045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-scripts\") pod \"22235e5d-c2cb-473f-a7aa-06f4ff712608\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.971062 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"22235e5d-c2cb-473f-a7aa-06f4ff712608\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.971114 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-logs\") pod \"22235e5d-c2cb-473f-a7aa-06f4ff712608\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.971201 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-combined-ca-bundle\") pod \"22235e5d-c2cb-473f-a7aa-06f4ff712608\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.971236 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-httpd-run\") pod \"22235e5d-c2cb-473f-a7aa-06f4ff712608\" (UID: \"22235e5d-c2cb-473f-a7aa-06f4ff712608\") " Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.973702 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "22235e5d-c2cb-473f-a7aa-06f4ff712608" (UID: "22235e5d-c2cb-473f-a7aa-06f4ff712608"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.973971 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-logs" (OuterVolumeSpecName: "logs") pod "22235e5d-c2cb-473f-a7aa-06f4ff712608" (UID: "22235e5d-c2cb-473f-a7aa-06f4ff712608"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.976300 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-scripts" (OuterVolumeSpecName: "scripts") pod "22235e5d-c2cb-473f-a7aa-06f4ff712608" (UID: "22235e5d-c2cb-473f-a7aa-06f4ff712608"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.980603 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "22235e5d-c2cb-473f-a7aa-06f4ff712608" (UID: "22235e5d-c2cb-473f-a7aa-06f4ff712608"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:35:12 crc kubenswrapper[4744]: I1201 08:35:12.985724 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22235e5d-c2cb-473f-a7aa-06f4ff712608-kube-api-access-49hds" (OuterVolumeSpecName: "kube-api-access-49hds") pod "22235e5d-c2cb-473f-a7aa-06f4ff712608" (UID: "22235e5d-c2cb-473f-a7aa-06f4ff712608"). InnerVolumeSpecName "kube-api-access-49hds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.008802 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22235e5d-c2cb-473f-a7aa-06f4ff712608" (UID: "22235e5d-c2cb-473f-a7aa-06f4ff712608"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.040242 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-config-data" (OuterVolumeSpecName: "config-data") pod "22235e5d-c2cb-473f-a7aa-06f4ff712608" (UID: "22235e5d-c2cb-473f-a7aa-06f4ff712608"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.073199 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.073224 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.073232 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.073240 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49hds\" (UniqueName: \"kubernetes.io/projected/22235e5d-c2cb-473f-a7aa-06f4ff712608-kube-api-access-49hds\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.073249 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22235e5d-c2cb-473f-a7aa-06f4ff712608-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.073278 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.073286 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22235e5d-c2cb-473f-a7aa-06f4ff712608-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.096469 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.174425 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.314988 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22235e5d-c2cb-473f-a7aa-06f4ff712608","Type":"ContainerDied","Data":"985a083bd3f10cf2b97cecea7493e621d55e45ce81330f3126de11d34e9d84dd"} Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.315037 4744 scope.go:117] "RemoveContainer" containerID="60c9039618c0dce314616a46e3c1aac7ee9e8543ccc804a91bf68e1160939c6f" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.315138 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.351887 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.367618 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.378812 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:13 crc kubenswrapper[4744]: E1201 08:35:13.379216 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-httpd" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.379236 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-httpd" Dec 01 08:35:13 crc kubenswrapper[4744]: E1201 08:35:13.379273 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-log" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.379282 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-log" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.379523 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-log" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.379562 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" containerName="glance-httpd" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.380511 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.387514 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.400971 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.441771 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.483482 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.483565 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.483597 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.483991 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.484042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.484088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mbx4\" (UniqueName: \"kubernetes.io/projected/1b0c98eb-08c5-4c49-b88b-8bb17594651e-kube-api-access-6mbx4\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.484120 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.584971 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.585029 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.585070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.585089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.585128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mbx4\" (UniqueName: \"kubernetes.io/projected/1b0c98eb-08c5-4c49-b88b-8bb17594651e-kube-api-access-6mbx4\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.585207 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.585926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.585994 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.586517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.587827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.590186 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.590680 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.599989 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.606133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mbx4\" (UniqueName: \"kubernetes.io/projected/1b0c98eb-08c5-4c49-b88b-8bb17594651e-kube-api-access-6mbx4\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.622132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: I1201 08:35:13.712705 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:13 crc kubenswrapper[4744]: E1201 08:35:13.890385 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6556f94_84b8_4cbc_b93f_464df38b5a15.slice/crio-b92d0cd7d77d5e49e122180332b578460f2a07dc39bfee62e14671c8a66b2d78.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6556f94_84b8_4cbc_b93f_464df38b5a15.slice/crio-conmon-b92d0cd7d77d5e49e122180332b578460f2a07dc39bfee62e14671c8a66b2d78.scope\": RecentStats: unable to find data in memory cache]" Dec 01 08:35:14 crc kubenswrapper[4744]: I1201 08:35:14.302041 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22235e5d-c2cb-473f-a7aa-06f4ff712608" path="/var/lib/kubelet/pods/22235e5d-c2cb-473f-a7aa-06f4ff712608/volumes" Dec 01 08:35:14 crc kubenswrapper[4744]: I1201 08:35:14.302909 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9efef428-22bb-47d2-a871-18cdc5495886" path="/var/lib/kubelet/pods/9efef428-22bb-47d2-a871-18cdc5495886/volumes" Dec 01 08:35:14 crc kubenswrapper[4744]: I1201 08:35:14.304817 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bce6df6c-65f0-4984-8887-4378ec0b59ea" path="/var/lib/kubelet/pods/bce6df6c-65f0-4984-8887-4378ec0b59ea/volumes" Dec 01 08:35:14 crc kubenswrapper[4744]: I1201 08:35:14.339287 4744 generic.go:334] "Generic (PLEG): container finished" podID="c6556f94-84b8-4cbc-b93f-464df38b5a15" containerID="b92d0cd7d77d5e49e122180332b578460f2a07dc39bfee62e14671c8a66b2d78" exitCode=0 Dec 01 08:35:14 crc kubenswrapper[4744]: I1201 08:35:14.339334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhv7l" event={"ID":"c6556f94-84b8-4cbc-b93f-464df38b5a15","Type":"ContainerDied","Data":"b92d0cd7d77d5e49e122180332b578460f2a07dc39bfee62e14671c8a66b2d78"} Dec 01 08:35:15 crc kubenswrapper[4744]: I1201 08:35:15.717734 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:15 crc kubenswrapper[4744]: I1201 08:35:15.780447 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:16 crc kubenswrapper[4744]: I1201 08:35:16.404635 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:16 crc kubenswrapper[4744]: I1201 08:35:16.451371 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-7csls"] Dec 01 08:35:16 crc kubenswrapper[4744]: I1201 08:35:16.451724 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="dnsmasq-dns" containerID="cri-o://fdcc8814fe3a8144d45af81809fe8b51fb45c66b00bff436699248088a1eb92b" gracePeriod=10 Dec 01 08:35:17 crc kubenswrapper[4744]: I1201 08:35:17.373158 4744 generic.go:334] "Generic (PLEG): container finished" podID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerID="fdcc8814fe3a8144d45af81809fe8b51fb45c66b00bff436699248088a1eb92b" exitCode=0 Dec 01 08:35:17 crc kubenswrapper[4744]: I1201 08:35:17.373193 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" event={"ID":"68fd39e9-b7c3-42db-8fcb-a0b39f586693","Type":"ContainerDied","Data":"fdcc8814fe3a8144d45af81809fe8b51fb45c66b00bff436699248088a1eb92b"} Dec 01 08:35:18 crc kubenswrapper[4744]: I1201 08:35:18.875694 4744 scope.go:117] "RemoveContainer" containerID="ae6589dc2f4b0ddedb8a62ece74df90a775d331578f6c67d02ca20c3228d0db3" Dec 01 08:35:18 crc kubenswrapper[4744]: I1201 08:35:18.916917 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.089007 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qgc8\" (UniqueName: \"kubernetes.io/projected/c6556f94-84b8-4cbc-b93f-464df38b5a15-kube-api-access-2qgc8\") pod \"c6556f94-84b8-4cbc-b93f-464df38b5a15\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.089125 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-combined-ca-bundle\") pod \"c6556f94-84b8-4cbc-b93f-464df38b5a15\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.089178 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-credential-keys\") pod \"c6556f94-84b8-4cbc-b93f-464df38b5a15\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.089237 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-scripts\") pod \"c6556f94-84b8-4cbc-b93f-464df38b5a15\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.089263 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-config-data\") pod \"c6556f94-84b8-4cbc-b93f-464df38b5a15\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.089291 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-fernet-keys\") pod \"c6556f94-84b8-4cbc-b93f-464df38b5a15\" (UID: \"c6556f94-84b8-4cbc-b93f-464df38b5a15\") " Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.095910 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6556f94-84b8-4cbc-b93f-464df38b5a15-kube-api-access-2qgc8" (OuterVolumeSpecName: "kube-api-access-2qgc8") pod "c6556f94-84b8-4cbc-b93f-464df38b5a15" (UID: "c6556f94-84b8-4cbc-b93f-464df38b5a15"). InnerVolumeSpecName "kube-api-access-2qgc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.097536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c6556f94-84b8-4cbc-b93f-464df38b5a15" (UID: "c6556f94-84b8-4cbc-b93f-464df38b5a15"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.102128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-scripts" (OuterVolumeSpecName: "scripts") pod "c6556f94-84b8-4cbc-b93f-464df38b5a15" (UID: "c6556f94-84b8-4cbc-b93f-464df38b5a15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.113070 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c6556f94-84b8-4cbc-b93f-464df38b5a15" (UID: "c6556f94-84b8-4cbc-b93f-464df38b5a15"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.120748 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6556f94-84b8-4cbc-b93f-464df38b5a15" (UID: "c6556f94-84b8-4cbc-b93f-464df38b5a15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.131238 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-config-data" (OuterVolumeSpecName: "config-data") pod "c6556f94-84b8-4cbc-b93f-464df38b5a15" (UID: "c6556f94-84b8-4cbc-b93f-464df38b5a15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.191375 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.191432 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.191447 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.191457 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.191467 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6556f94-84b8-4cbc-b93f-464df38b5a15-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.191477 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qgc8\" (UniqueName: \"kubernetes.io/projected/c6556f94-84b8-4cbc-b93f-464df38b5a15-kube-api-access-2qgc8\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.393241 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"abf48605-5626-4522-bab4-63b0ff51943a","Type":"ContainerStarted","Data":"7cb53e83468abeb5cb53e1c4d7b641e1868f07ebd4731d8dfeface172210dfd1"} Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.398040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nhv7l" event={"ID":"c6556f94-84b8-4cbc-b93f-464df38b5a15","Type":"ContainerDied","Data":"c37179e9d623fbdf9d2a1c3e5c8c4cccf1178b476430539fafa1694c48e24820"} Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.398068 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c37179e9d623fbdf9d2a1c3e5c8c4cccf1178b476430539fafa1694c48e24820" Dec 01 08:35:19 crc kubenswrapper[4744]: I1201 08:35:19.398129 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nhv7l" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.013965 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nhv7l"] Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.023765 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nhv7l"] Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.090350 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xdnlb"] Dec 01 08:35:20 crc kubenswrapper[4744]: E1201 08:35:20.090717 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6556f94-84b8-4cbc-b93f-464df38b5a15" containerName="keystone-bootstrap" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.090734 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6556f94-84b8-4cbc-b93f-464df38b5a15" containerName="keystone-bootstrap" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.090887 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6556f94-84b8-4cbc-b93f-464df38b5a15" containerName="keystone-bootstrap" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.091468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.094099 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.094292 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.094844 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4jgn6" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.094884 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.095037 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.099785 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xdnlb"] Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.206454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-combined-ca-bundle\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.206752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-fernet-keys\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.206818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-credential-keys\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.206887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-scripts\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.206905 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7krvk\" (UniqueName: \"kubernetes.io/projected/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-kube-api-access-7krvk\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.206923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-config-data\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.302070 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6556f94-84b8-4cbc-b93f-464df38b5a15" path="/var/lib/kubelet/pods/c6556f94-84b8-4cbc-b93f-464df38b5a15/volumes" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.309533 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-credential-keys\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.310864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-scripts\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.310997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7krvk\" (UniqueName: \"kubernetes.io/projected/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-kube-api-access-7krvk\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.311054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-config-data\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.311225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-combined-ca-bundle\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.311288 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-fernet-keys\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.322634 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-combined-ca-bundle\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.323121 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-config-data\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.323245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-fernet-keys\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.324780 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-scripts\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.325337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-credential-keys\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.327513 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7krvk\" (UniqueName: \"kubernetes.io/projected/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-kube-api-access-7krvk\") pod \"keystone-bootstrap-xdnlb\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:20 crc kubenswrapper[4744]: I1201 08:35:20.460292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:25 crc kubenswrapper[4744]: I1201 08:35:25.093633 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.025896 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.143911 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s6cx\" (UniqueName: \"kubernetes.io/projected/68fd39e9-b7c3-42db-8fcb-a0b39f586693-kube-api-access-7s6cx\") pod \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.144019 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-nb\") pod \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.144045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-sb\") pod \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.144086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-svc\") pod \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.144128 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-swift-storage-0\") pod \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.144174 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-config\") pod \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\" (UID: \"68fd39e9-b7c3-42db-8fcb-a0b39f586693\") " Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.149703 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fd39e9-b7c3-42db-8fcb-a0b39f586693-kube-api-access-7s6cx" (OuterVolumeSpecName: "kube-api-access-7s6cx") pod "68fd39e9-b7c3-42db-8fcb-a0b39f586693" (UID: "68fd39e9-b7c3-42db-8fcb-a0b39f586693"). InnerVolumeSpecName "kube-api-access-7s6cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.190614 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-config" (OuterVolumeSpecName: "config") pod "68fd39e9-b7c3-42db-8fcb-a0b39f586693" (UID: "68fd39e9-b7c3-42db-8fcb-a0b39f586693"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.191248 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68fd39e9-b7c3-42db-8fcb-a0b39f586693" (UID: "68fd39e9-b7c3-42db-8fcb-a0b39f586693"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.207212 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68fd39e9-b7c3-42db-8fcb-a0b39f586693" (UID: "68fd39e9-b7c3-42db-8fcb-a0b39f586693"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.208757 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "68fd39e9-b7c3-42db-8fcb-a0b39f586693" (UID: "68fd39e9-b7c3-42db-8fcb-a0b39f586693"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.225844 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "68fd39e9-b7c3-42db-8fcb-a0b39f586693" (UID: "68fd39e9-b7c3-42db-8fcb-a0b39f586693"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.245892 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.245927 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s6cx\" (UniqueName: \"kubernetes.io/projected/68fd39e9-b7c3-42db-8fcb-a0b39f586693-kube-api-access-7s6cx\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.245940 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.245951 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.245962 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.245972 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68fd39e9-b7c3-42db-8fcb-a0b39f586693-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.453787 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.478650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" event={"ID":"68fd39e9-b7c3-42db-8fcb-a0b39f586693","Type":"ContainerDied","Data":"1e8a3f4beb49a113618e2e1b1ea76e0fd75e4db640c6098aa0eaf5abbbedbbc9"} Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.478710 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.514521 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-7csls"] Dec 01 08:35:27 crc kubenswrapper[4744]: I1201 08:35:27.522021 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-7csls"] Dec 01 08:35:28 crc kubenswrapper[4744]: I1201 08:35:28.293575 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" path="/var/lib/kubelet/pods/68fd39e9-b7c3-42db-8fcb-a0b39f586693/volumes" Dec 01 08:35:30 crc kubenswrapper[4744]: I1201 08:35:30.095135 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-7csls" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 01 08:35:31 crc kubenswrapper[4744]: I1201 08:35:31.514867 4744 generic.go:334] "Generic (PLEG): container finished" podID="d51f040d-3a61-47f1-8699-c88fbaabdee5" containerID="3c4190b993a15f411262d140d6dc55defae79c915dc8bbfafcb13a406c81c171" exitCode=0 Dec 01 08:35:31 crc kubenswrapper[4744]: I1201 08:35:31.514951 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5k8xk" event={"ID":"d51f040d-3a61-47f1-8699-c88fbaabdee5","Type":"ContainerDied","Data":"3c4190b993a15f411262d140d6dc55defae79c915dc8bbfafcb13a406c81c171"} Dec 01 08:35:35 crc kubenswrapper[4744]: E1201 08:35:35.487302 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 01 08:35:35 crc kubenswrapper[4744]: E1201 08:35:35.488109 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4nj56,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-x664l_openstack(ab3b54ed-348a-4314-8e92-3e977ee8f1ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:35:35 crc kubenswrapper[4744]: E1201 08:35:35.489473 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-x664l" podUID="ab3b54ed-348a-4314-8e92-3e977ee8f1ff" Dec 01 08:35:35 crc kubenswrapper[4744]: E1201 08:35:35.561684 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-x664l" podUID="ab3b54ed-348a-4314-8e92-3e977ee8f1ff" Dec 01 08:35:36 crc kubenswrapper[4744]: E1201 08:35:36.076985 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 01 08:35:36 crc kubenswrapper[4744]: E1201 08:35:36.077119 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m9jth,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-8mwhk_openstack(ed2a24ae-749f-4faf-b077-060df14120bb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:35:36 crc kubenswrapper[4744]: E1201 08:35:36.079170 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-8mwhk" podUID="ed2a24ae-749f-4faf-b077-060df14120bb" Dec 01 08:35:36 crc kubenswrapper[4744]: I1201 08:35:36.083786 4744 scope.go:117] "RemoveContainer" containerID="fdcc8814fe3a8144d45af81809fe8b51fb45c66b00bff436699248088a1eb92b" Dec 01 08:35:36 crc kubenswrapper[4744]: I1201 08:35:36.575821 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b0c98eb-08c5-4c49-b88b-8bb17594651e","Type":"ContainerStarted","Data":"0ab7212161fcbb334c6ff0900acee560f682fd94fdfe09a6a2a3a6f9c6e8e29f"} Dec 01 08:35:36 crc kubenswrapper[4744]: E1201 08:35:36.579805 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-8mwhk" podUID="ed2a24ae-749f-4faf-b077-060df14120bb" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.314824 4744 scope.go:117] "RemoveContainer" containerID="dcf876e7115e0164cd133ed8d7f5180a7f7826bc99870fff34bf83aec915f4c7" Dec 01 08:35:37 crc kubenswrapper[4744]: E1201 08:35:37.350767 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 01 08:35:37 crc kubenswrapper[4744]: E1201 08:35:37.350920 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv46w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-t2mxc_openstack(d3718521-a22d-4350-aa32-98987bfe5632): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 08:35:37 crc kubenswrapper[4744]: E1201 08:35:37.352290 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-t2mxc" podUID="d3718521-a22d-4350-aa32-98987bfe5632" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.545443 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.565312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq2xc\" (UniqueName: \"kubernetes.io/projected/d51f040d-3a61-47f1-8699-c88fbaabdee5-kube-api-access-jq2xc\") pod \"d51f040d-3a61-47f1-8699-c88fbaabdee5\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.565378 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-combined-ca-bundle\") pod \"d51f040d-3a61-47f1-8699-c88fbaabdee5\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.565557 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-config\") pod \"d51f040d-3a61-47f1-8699-c88fbaabdee5\" (UID: \"d51f040d-3a61-47f1-8699-c88fbaabdee5\") " Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.580185 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d51f040d-3a61-47f1-8699-c88fbaabdee5-kube-api-access-jq2xc" (OuterVolumeSpecName: "kube-api-access-jq2xc") pod "d51f040d-3a61-47f1-8699-c88fbaabdee5" (UID: "d51f040d-3a61-47f1-8699-c88fbaabdee5"). InnerVolumeSpecName "kube-api-access-jq2xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.589725 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5k8xk" event={"ID":"d51f040d-3a61-47f1-8699-c88fbaabdee5","Type":"ContainerDied","Data":"48a3d190ed79a2c929bbf02730fa7b78396a4cdc7c011028b689198155f76f1c"} Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.589761 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48a3d190ed79a2c929bbf02730fa7b78396a4cdc7c011028b689198155f76f1c" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.589814 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5k8xk" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.598718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-config" (OuterVolumeSpecName: "config") pod "d51f040d-3a61-47f1-8699-c88fbaabdee5" (UID: "d51f040d-3a61-47f1-8699-c88fbaabdee5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:37 crc kubenswrapper[4744]: E1201 08:35:37.608225 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-t2mxc" podUID="d3718521-a22d-4350-aa32-98987bfe5632" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.635822 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d51f040d-3a61-47f1-8699-c88fbaabdee5" (UID: "d51f040d-3a61-47f1-8699-c88fbaabdee5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.666705 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq2xc\" (UniqueName: \"kubernetes.io/projected/d51f040d-3a61-47f1-8699-c88fbaabdee5-kube-api-access-jq2xc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.666733 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.666742 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d51f040d-3a61-47f1-8699-c88fbaabdee5-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:37 crc kubenswrapper[4744]: I1201 08:35:37.785448 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xdnlb"] Dec 01 08:35:37 crc kubenswrapper[4744]: W1201 08:35:37.791167 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf20b7a9_6976_4f37_8cfc_f3c67d2057c2.slice/crio-4b63c38ac0adbb05c96f5bf132cfb3403c3c6d2c663f7720b224ef36a3879403 WatchSource:0}: Error finding container 4b63c38ac0adbb05c96f5bf132cfb3403c3c6d2c663f7720b224ef36a3879403: Status 404 returned error can't find the container with id 4b63c38ac0adbb05c96f5bf132cfb3403c3c6d2c663f7720b224ef36a3879403 Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.605138 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerStarted","Data":"179e72728e42caf2adc1f8f305005ce810fc06276b7326ef6ea701ce3d37c130"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.607664 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"abf48605-5626-4522-bab4-63b0ff51943a","Type":"ContainerStarted","Data":"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.607739 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"abf48605-5626-4522-bab4-63b0ff51943a","Type":"ContainerStarted","Data":"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.607893 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-log" containerID="cri-o://769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678" gracePeriod=30 Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.608444 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-httpd" containerID="cri-o://9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95" gracePeriod=30 Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.610787 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bzxk9" event={"ID":"95085c33-3a8b-4af5-aaef-bd718dabcf6f","Type":"ContainerStarted","Data":"803acab525a520838fffbec016e53bdc2404fe4cd412bcc287c30998b04dc8dd"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.617038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdnlb" event={"ID":"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2","Type":"ContainerStarted","Data":"cf486ae0b0a4c2c372ea294f78931e59ebf5398ab8cd5b7a34c4a3a403a4032a"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.617087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdnlb" event={"ID":"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2","Type":"ContainerStarted","Data":"4b63c38ac0adbb05c96f5bf132cfb3403c3c6d2c663f7720b224ef36a3879403"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.619881 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b0c98eb-08c5-4c49-b88b-8bb17594651e","Type":"ContainerStarted","Data":"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.619914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b0c98eb-08c5-4c49-b88b-8bb17594651e","Type":"ContainerStarted","Data":"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541"} Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.620039 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-log" containerID="cri-o://66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541" gracePeriod=30 Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.620101 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-httpd" containerID="cri-o://3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191" gracePeriod=30 Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.648053 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=26.648033351 podStartE2EDuration="26.648033351s" podCreationTimestamp="2025-12-01 08:35:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:38.644368806 +0000 UTC m=+1070.633426727" watchObservedRunningTime="2025-12-01 08:35:38.648033351 +0000 UTC m=+1070.637091282" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.666763 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xdnlb" podStartSLOduration=18.666748254 podStartE2EDuration="18.666748254s" podCreationTimestamp="2025-12-01 08:35:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:38.65925129 +0000 UTC m=+1070.648309211" watchObservedRunningTime="2025-12-01 08:35:38.666748254 +0000 UTC m=+1070.655806165" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.679211 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-bzxk9" podStartSLOduration=4.25545831 podStartE2EDuration="33.679193508s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="2025-12-01 08:35:07.874634005 +0000 UTC m=+1039.863691926" lastFinishedPulling="2025-12-01 08:35:37.298369183 +0000 UTC m=+1069.287427124" observedRunningTime="2025-12-01 08:35:38.673072064 +0000 UTC m=+1070.662129985" watchObservedRunningTime="2025-12-01 08:35:38.679193508 +0000 UTC m=+1070.668251429" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.693853 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=25.693832225 podStartE2EDuration="25.693832225s" podCreationTimestamp="2025-12-01 08:35:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:38.69225533 +0000 UTC m=+1070.681313261" watchObservedRunningTime="2025-12-01 08:35:38.693832225 +0000 UTC m=+1070.682890156" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.802861 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-pjtl9"] Dec 01 08:35:38 crc kubenswrapper[4744]: E1201 08:35:38.803305 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d51f040d-3a61-47f1-8699-c88fbaabdee5" containerName="neutron-db-sync" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.803321 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d51f040d-3a61-47f1-8699-c88fbaabdee5" containerName="neutron-db-sync" Dec 01 08:35:38 crc kubenswrapper[4744]: E1201 08:35:38.803365 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="init" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.803374 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="init" Dec 01 08:35:38 crc kubenswrapper[4744]: E1201 08:35:38.803417 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="dnsmasq-dns" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.803427 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="dnsmasq-dns" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.803653 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fd39e9-b7c3-42db-8fcb-a0b39f586693" containerName="dnsmasq-dns" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.803673 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d51f040d-3a61-47f1-8699-c88fbaabdee5" containerName="neutron-db-sync" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.805220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.825994 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-pjtl9"] Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.907119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.907162 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.907212 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.907232 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.907293 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-224d8\" (UniqueName: \"kubernetes.io/projected/3dfafea4-b9db-479a-9876-714ee1dfb14b-kube-api-access-224d8\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.907332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-config\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.940638 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-84cf778f56-stb8l"] Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.942346 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.945696 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.945716 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.945889 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4wwpn" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.946469 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 08:35:38 crc kubenswrapper[4744]: I1201 08:35:38.953497 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84cf778f56-stb8l"] Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.009182 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.009230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.009287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.009316 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.009363 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-224d8\" (UniqueName: \"kubernetes.io/projected/3dfafea4-b9db-479a-9876-714ee1dfb14b-kube-api-access-224d8\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.009431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-config\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.010123 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.010527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-config\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.010846 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.011312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.011721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.033611 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-224d8\" (UniqueName: \"kubernetes.io/projected/3dfafea4-b9db-479a-9876-714ee1dfb14b-kube-api-access-224d8\") pod \"dnsmasq-dns-5ccc5c4795-pjtl9\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.111103 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-combined-ca-bundle\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.111532 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x6jn\" (UniqueName: \"kubernetes.io/projected/39694e5c-5d4d-4d84-803b-1557e0c49f43-kube-api-access-9x6jn\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.111662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-ovndb-tls-certs\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.111808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-config\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.111940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-httpd-config\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.199737 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.213288 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-config\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.213335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-httpd-config\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.213400 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-combined-ca-bundle\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.213453 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x6jn\" (UniqueName: \"kubernetes.io/projected/39694e5c-5d4d-4d84-803b-1557e0c49f43-kube-api-access-9x6jn\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.213484 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-ovndb-tls-certs\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.220132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-ovndb-tls-certs\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.220940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-httpd-config\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.225781 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-combined-ca-bundle\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.230179 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-config\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.268674 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x6jn\" (UniqueName: \"kubernetes.io/projected/39694e5c-5d4d-4d84-803b-1557e0c49f43-kube-api-access-9x6jn\") pod \"neutron-84cf778f56-stb8l\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.271919 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.530778 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.570160 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.632228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"abf48605-5626-4522-bab4-63b0ff51943a\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.632265 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-logs\") pod \"abf48605-5626-4522-bab4-63b0ff51943a\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.632321 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-scripts\") pod \"abf48605-5626-4522-bab4-63b0ff51943a\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.632350 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2tpq\" (UniqueName: \"kubernetes.io/projected/abf48605-5626-4522-bab4-63b0ff51943a-kube-api-access-h2tpq\") pod \"abf48605-5626-4522-bab4-63b0ff51943a\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.632384 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-combined-ca-bundle\") pod \"abf48605-5626-4522-bab4-63b0ff51943a\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.632427 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-config-data\") pod \"abf48605-5626-4522-bab4-63b0ff51943a\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.632512 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-httpd-run\") pod \"abf48605-5626-4522-bab4-63b0ff51943a\" (UID: \"abf48605-5626-4522-bab4-63b0ff51943a\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.633379 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "abf48605-5626-4522-bab4-63b0ff51943a" (UID: "abf48605-5626-4522-bab4-63b0ff51943a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.634111 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-logs" (OuterVolumeSpecName: "logs") pod "abf48605-5626-4522-bab4-63b0ff51943a" (UID: "abf48605-5626-4522-bab4-63b0ff51943a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.640717 4744 generic.go:334] "Generic (PLEG): container finished" podID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerID="3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191" exitCode=143 Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.640758 4744 generic.go:334] "Generic (PLEG): container finished" podID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerID="66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541" exitCode=143 Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.640772 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b0c98eb-08c5-4c49-b88b-8bb17594651e","Type":"ContainerDied","Data":"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191"} Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.640821 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b0c98eb-08c5-4c49-b88b-8bb17594651e","Type":"ContainerDied","Data":"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541"} Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.640833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1b0c98eb-08c5-4c49-b88b-8bb17594651e","Type":"ContainerDied","Data":"0ab7212161fcbb334c6ff0900acee560f682fd94fdfe09a6a2a3a6f9c6e8e29f"} Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.640851 4744 scope.go:117] "RemoveContainer" containerID="3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.640962 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.644104 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-scripts" (OuterVolumeSpecName: "scripts") pod "abf48605-5626-4522-bab4-63b0ff51943a" (UID: "abf48605-5626-4522-bab4-63b0ff51943a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.644222 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf48605-5626-4522-bab4-63b0ff51943a-kube-api-access-h2tpq" (OuterVolumeSpecName: "kube-api-access-h2tpq") pod "abf48605-5626-4522-bab4-63b0ff51943a" (UID: "abf48605-5626-4522-bab4-63b0ff51943a"). InnerVolumeSpecName "kube-api-access-h2tpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.648732 4744 generic.go:334] "Generic (PLEG): container finished" podID="abf48605-5626-4522-bab4-63b0ff51943a" containerID="9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95" exitCode=143 Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.648759 4744 generic.go:334] "Generic (PLEG): container finished" podID="abf48605-5626-4522-bab4-63b0ff51943a" containerID="769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678" exitCode=143 Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.648792 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.648793 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"abf48605-5626-4522-bab4-63b0ff51943a","Type":"ContainerDied","Data":"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95"} Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.648841 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"abf48605-5626-4522-bab4-63b0ff51943a","Type":"ContainerDied","Data":"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678"} Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.648858 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"abf48605-5626-4522-bab4-63b0ff51943a","Type":"ContainerDied","Data":"7cb53e83468abeb5cb53e1c4d7b641e1868f07ebd4731d8dfeface172210dfd1"} Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.650163 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "abf48605-5626-4522-bab4-63b0ff51943a" (UID: "abf48605-5626-4522-bab4-63b0ff51943a"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.697590 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abf48605-5626-4522-bab4-63b0ff51943a" (UID: "abf48605-5626-4522-bab4-63b0ff51943a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mbx4\" (UniqueName: \"kubernetes.io/projected/1b0c98eb-08c5-4c49-b88b-8bb17594651e-kube-api-access-6mbx4\") pod \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737232 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-config-data\") pod \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-httpd-run\") pod \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737369 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-logs\") pod \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737422 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-scripts\") pod \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737479 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-combined-ca-bundle\") pod \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\" (UID: \"1b0c98eb-08c5-4c49-b88b-8bb17594651e\") " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737891 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737907 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737924 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737933 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abf48605-5626-4522-bab4-63b0ff51943a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737941 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.737949 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2tpq\" (UniqueName: \"kubernetes.io/projected/abf48605-5626-4522-bab4-63b0ff51943a-kube-api-access-h2tpq\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.738897 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-logs" (OuterVolumeSpecName: "logs") pod "1b0c98eb-08c5-4c49-b88b-8bb17594651e" (UID: "1b0c98eb-08c5-4c49-b88b-8bb17594651e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.741445 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1b0c98eb-08c5-4c49-b88b-8bb17594651e" (UID: "1b0c98eb-08c5-4c49-b88b-8bb17594651e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.750959 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b0c98eb-08c5-4c49-b88b-8bb17594651e-kube-api-access-6mbx4" (OuterVolumeSpecName: "kube-api-access-6mbx4") pod "1b0c98eb-08c5-4c49-b88b-8bb17594651e" (UID: "1b0c98eb-08c5-4c49-b88b-8bb17594651e"). InnerVolumeSpecName "kube-api-access-6mbx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.751037 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "1b0c98eb-08c5-4c49-b88b-8bb17594651e" (UID: "1b0c98eb-08c5-4c49-b88b-8bb17594651e"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.770916 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-scripts" (OuterVolumeSpecName: "scripts") pod "1b0c98eb-08c5-4c49-b88b-8bb17594651e" (UID: "1b0c98eb-08c5-4c49-b88b-8bb17594651e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.776882 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-config-data" (OuterVolumeSpecName: "config-data") pod "abf48605-5626-4522-bab4-63b0ff51943a" (UID: "abf48605-5626-4522-bab4-63b0ff51943a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.795115 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.842850 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mbx4\" (UniqueName: \"kubernetes.io/projected/1b0c98eb-08c5-4c49-b88b-8bb17594651e-kube-api-access-6mbx4\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.843246 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.843335 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.843398 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.843483 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0c98eb-08c5-4c49-b88b-8bb17594651e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.843544 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.843602 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abf48605-5626-4522-bab4-63b0ff51943a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.891138 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.917746 4744 scope.go:117] "RemoveContainer" containerID="66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.918195 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b0c98eb-08c5-4c49-b88b-8bb17594651e" (UID: "1b0c98eb-08c5-4c49-b88b-8bb17594651e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.935051 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-config-data" (OuterVolumeSpecName: "config-data") pod "1b0c98eb-08c5-4c49-b88b-8bb17594651e" (UID: "1b0c98eb-08c5-4c49-b88b-8bb17594651e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.949682 4744 scope.go:117] "RemoveContainer" containerID="3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191" Dec 01 08:35:39 crc kubenswrapper[4744]: E1201 08:35:39.951368 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191\": container with ID starting with 3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191 not found: ID does not exist" containerID="3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.951521 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191"} err="failed to get container status \"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191\": rpc error: code = NotFound desc = could not find container \"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191\": container with ID starting with 3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191 not found: ID does not exist" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.951660 4744 scope.go:117] "RemoveContainer" containerID="66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541" Dec 01 08:35:39 crc kubenswrapper[4744]: E1201 08:35:39.951859 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541\": container with ID starting with 66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541 not found: ID does not exist" containerID="66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.951885 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541"} err="failed to get container status \"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541\": rpc error: code = NotFound desc = could not find container \"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541\": container with ID starting with 66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541 not found: ID does not exist" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.951899 4744 scope.go:117] "RemoveContainer" containerID="3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.952116 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191"} err="failed to get container status \"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191\": rpc error: code = NotFound desc = could not find container \"3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191\": container with ID starting with 3eb7025320fa24e076f95d7643f5a1d38fa0316e3304809c2d2fff4842a8c191 not found: ID does not exist" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.952134 4744 scope.go:117] "RemoveContainer" containerID="66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.952311 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541"} err="failed to get container status \"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541\": rpc error: code = NotFound desc = could not find container \"66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541\": container with ID starting with 66542f281d20675852024dce82b46c9571be4306fa1f956bb22e573905e64541 not found: ID does not exist" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.952327 4744 scope.go:117] "RemoveContainer" containerID="9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.956471 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.957071 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b0c98eb-08c5-4c49-b88b-8bb17594651e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.957116 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.958626 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-pjtl9"] Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.986009 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84cf778f56-stb8l"] Dec 01 08:35:39 crc kubenswrapper[4744]: I1201 08:35:39.994451 4744 scope.go:117] "RemoveContainer" containerID="769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.010679 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.020900 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.046089 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.060100 4744 scope.go:117] "RemoveContainer" containerID="9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95" Dec 01 08:35:40 crc kubenswrapper[4744]: E1201 08:35:40.061192 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95\": container with ID starting with 9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95 not found: ID does not exist" containerID="9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.061224 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95"} err="failed to get container status \"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95\": rpc error: code = NotFound desc = could not find container \"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95\": container with ID starting with 9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95 not found: ID does not exist" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.061245 4744 scope.go:117] "RemoveContainer" containerID="769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678" Dec 01 08:35:40 crc kubenswrapper[4744]: E1201 08:35:40.061573 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678\": container with ID starting with 769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678 not found: ID does not exist" containerID="769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.061594 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678"} err="failed to get container status \"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678\": rpc error: code = NotFound desc = could not find container \"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678\": container with ID starting with 769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678 not found: ID does not exist" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.061607 4744 scope.go:117] "RemoveContainer" containerID="9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.062052 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95"} err="failed to get container status \"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95\": rpc error: code = NotFound desc = could not find container \"9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95\": container with ID starting with 9f85e0ed4cb7aa79f726f9f11199c190523adef9a858d4c7f3bd54eb7b332a95 not found: ID does not exist" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.062118 4744 scope.go:117] "RemoveContainer" containerID="769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.062399 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678"} err="failed to get container status \"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678\": rpc error: code = NotFound desc = could not find container \"769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678\": container with ID starting with 769d2ee3549d078150f4624cdf093b7dcdfbd27c5cab80d8a9d7f38725318678 not found: ID does not exist" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.089440 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: E1201 08:35:40.089992 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-log" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090003 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-log" Dec 01 08:35:40 crc kubenswrapper[4744]: E1201 08:35:40.090015 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-httpd" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090021 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-httpd" Dec 01 08:35:40 crc kubenswrapper[4744]: E1201 08:35:40.090048 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-httpd" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090054 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-httpd" Dec 01 08:35:40 crc kubenswrapper[4744]: E1201 08:35:40.090064 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-log" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090072 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-log" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090253 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-log" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090265 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-httpd" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090274 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf48605-5626-4522-bab4-63b0ff51943a" containerName="glance-httpd" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.090289 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" containerName="glance-log" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.091211 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.107276 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.107483 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gl4sq" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.107591 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.107692 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.125629 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.144910 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.146358 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.148740 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.148884 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.159097 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.170718 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.262706 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-scripts\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.262830 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.262899 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.262920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.262949 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzng5\" (UniqueName: \"kubernetes.io/projected/4a0a609f-e643-4e6b-b0f8-428fd792fce9-kube-api-access-fzng5\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.262977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.262998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-logs\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263022 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-logs\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263057 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263097 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-config-data\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263158 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263184 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263229 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntnbh\" (UniqueName: \"kubernetes.io/projected/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-kube-api-access-ntnbh\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.263255 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.298245 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b0c98eb-08c5-4c49-b88b-8bb17594651e" path="/var/lib/kubelet/pods/1b0c98eb-08c5-4c49-b88b-8bb17594651e/volumes" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.298987 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf48605-5626-4522-bab4-63b0ff51943a" path="/var/lib/kubelet/pods/abf48605-5626-4522-bab4-63b0ff51943a/volumes" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364826 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-config-data\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364872 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364889 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntnbh\" (UniqueName: \"kubernetes.io/projected/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-kube-api-access-ntnbh\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364972 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.364999 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-scripts\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365039 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365082 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365100 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365118 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzng5\" (UniqueName: \"kubernetes.io/projected/4a0a609f-e643-4e6b-b0f8-428fd792fce9-kube-api-access-fzng5\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365141 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365158 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-logs\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365178 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-logs\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.365761 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.366058 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.366086 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.366104 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.366466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-logs\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.366503 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-logs\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.370791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.385686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-config-data\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.385841 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-scripts\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.388099 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.388341 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.390329 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.391311 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.402601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzng5\" (UniqueName: \"kubernetes.io/projected/4a0a609f-e643-4e6b-b0f8-428fd792fce9-kube-api-access-fzng5\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.406396 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntnbh\" (UniqueName: \"kubernetes.io/projected/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-kube-api-access-ntnbh\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.410400 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.423106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.431877 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.595260 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.687016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cf778f56-stb8l" event={"ID":"39694e5c-5d4d-4d84-803b-1557e0c49f43","Type":"ContainerStarted","Data":"481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed"} Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.687066 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cf778f56-stb8l" event={"ID":"39694e5c-5d4d-4d84-803b-1557e0c49f43","Type":"ContainerStarted","Data":"822c2216940c509513776fab6e7a8be9d828a5084cdad1945db6d0cc7e1c86d4"} Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.690723 4744 generic.go:334] "Generic (PLEG): container finished" podID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerID="e3aaf2c7e38e8abea2025dc2c20bb09439a96aa7f14cd14295b1892504aaf946" exitCode=0 Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.690778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" event={"ID":"3dfafea4-b9db-479a-9876-714ee1dfb14b","Type":"ContainerDied","Data":"e3aaf2c7e38e8abea2025dc2c20bb09439a96aa7f14cd14295b1892504aaf946"} Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.690797 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" event={"ID":"3dfafea4-b9db-479a-9876-714ee1dfb14b","Type":"ContainerStarted","Data":"2d22950627f919a594cbd1634edebb58431eab4776a9834ff3c94e62a9d5c81f"} Dec 01 08:35:40 crc kubenswrapper[4744]: I1201 08:35:40.725994 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.315077 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-685cbf69b9-jrc9r"] Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.316875 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.320178 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.320325 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.369579 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685cbf69b9-jrc9r"] Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.389738 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.463555 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.490357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-combined-ca-bundle\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.490440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86phv\" (UniqueName: \"kubernetes.io/projected/09924952-c000-422a-b471-e738246ae161-kube-api-access-86phv\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.490479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-ovndb-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.490626 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-internal-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.490686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-config\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.490713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-httpd-config\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.490768 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-public-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.592736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-config\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.592793 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-httpd-config\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.592846 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-public-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.592888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-combined-ca-bundle\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.592905 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86phv\" (UniqueName: \"kubernetes.io/projected/09924952-c000-422a-b471-e738246ae161-kube-api-access-86phv\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.592923 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-ovndb-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.592970 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-internal-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.599122 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-public-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.600896 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-config\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.601520 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-ovndb-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.601551 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-httpd-config\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.606485 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-combined-ca-bundle\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.617043 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09924952-c000-422a-b471-e738246ae161-internal-tls-certs\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.636312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86phv\" (UniqueName: \"kubernetes.io/projected/09924952-c000-422a-b471-e738246ae161-kube-api-access-86phv\") pod \"neutron-685cbf69b9-jrc9r\" (UID: \"09924952-c000-422a-b471-e738246ae161\") " pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: W1201 08:35:41.639075 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a0a609f_e643_4e6b_b0f8_428fd792fce9.slice/crio-6c3f52aa05b946ca3ba4e7c1e13f59c75f94872788bff36a88ff53951d745b98 WatchSource:0}: Error finding container 6c3f52aa05b946ca3ba4e7c1e13f59c75f94872788bff36a88ff53951d745b98: Status 404 returned error can't find the container with id 6c3f52aa05b946ca3ba4e7c1e13f59c75f94872788bff36a88ff53951d745b98 Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.692302 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.715911 4744 generic.go:334] "Generic (PLEG): container finished" podID="cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" containerID="cf486ae0b0a4c2c372ea294f78931e59ebf5398ab8cd5b7a34c4a3a403a4032a" exitCode=0 Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.715970 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdnlb" event={"ID":"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2","Type":"ContainerDied","Data":"cf486ae0b0a4c2c372ea294f78931e59ebf5398ab8cd5b7a34c4a3a403a4032a"} Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.719848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" event={"ID":"3dfafea4-b9db-479a-9876-714ee1dfb14b","Type":"ContainerStarted","Data":"fad23623055fac729576a02472859863ce2d225d63941d7207a0100984ae31de"} Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.720326 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.723079 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1691c44a-f2ba-43eb-9ff3-a6a25b352d78","Type":"ContainerStarted","Data":"0d33f494f418f0023078563f0809b0ab21d55b14b3f06995bc9ba03d989746c5"} Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.724796 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cf778f56-stb8l" event={"ID":"39694e5c-5d4d-4d84-803b-1557e0c49f43","Type":"ContainerStarted","Data":"c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547"} Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.725224 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.726552 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a0a609f-e643-4e6b-b0f8-428fd792fce9","Type":"ContainerStarted","Data":"6c3f52aa05b946ca3ba4e7c1e13f59c75f94872788bff36a88ff53951d745b98"} Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.736749 4744 generic.go:334] "Generic (PLEG): container finished" podID="95085c33-3a8b-4af5-aaef-bd718dabcf6f" containerID="803acab525a520838fffbec016e53bdc2404fe4cd412bcc287c30998b04dc8dd" exitCode=0 Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.736791 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bzxk9" event={"ID":"95085c33-3a8b-4af5-aaef-bd718dabcf6f","Type":"ContainerDied","Data":"803acab525a520838fffbec016e53bdc2404fe4cd412bcc287c30998b04dc8dd"} Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.765290 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-84cf778f56-stb8l" podStartSLOduration=3.765272865 podStartE2EDuration="3.765272865s" podCreationTimestamp="2025-12-01 08:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:41.748906219 +0000 UTC m=+1073.737964140" watchObservedRunningTime="2025-12-01 08:35:41.765272865 +0000 UTC m=+1073.754330786" Dec 01 08:35:41 crc kubenswrapper[4744]: I1201 08:35:41.779855 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" podStartSLOduration=3.77983869 podStartE2EDuration="3.77983869s" podCreationTimestamp="2025-12-01 08:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:41.768398664 +0000 UTC m=+1073.757456585" watchObservedRunningTime="2025-12-01 08:35:41.77983869 +0000 UTC m=+1073.768896611" Dec 01 08:35:42 crc kubenswrapper[4744]: I1201 08:35:42.224154 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-685cbf69b9-jrc9r"] Dec 01 08:35:42 crc kubenswrapper[4744]: W1201 08:35:42.225131 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09924952_c000_422a_b471_e738246ae161.slice/crio-4bbbf5fdce710a2bedd279768fd2657a627a798cf948db321fe2a09160807ff7 WatchSource:0}: Error finding container 4bbbf5fdce710a2bedd279768fd2657a627a798cf948db321fe2a09160807ff7: Status 404 returned error can't find the container with id 4bbbf5fdce710a2bedd279768fd2657a627a798cf948db321fe2a09160807ff7 Dec 01 08:35:42 crc kubenswrapper[4744]: I1201 08:35:42.795638 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerStarted","Data":"41101381762b71bd360b63ab1e8962d90cd86d9fc2e3a55615047f712ab5c16b"} Dec 01 08:35:42 crc kubenswrapper[4744]: I1201 08:35:42.813572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1691c44a-f2ba-43eb-9ff3-a6a25b352d78","Type":"ContainerStarted","Data":"fa8f79b5a76b19113710fe7c2f739a769d63613b3da2f9df83d8cc48399ddd90"} Dec 01 08:35:42 crc kubenswrapper[4744]: I1201 08:35:42.824450 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685cbf69b9-jrc9r" event={"ID":"09924952-c000-422a-b471-e738246ae161","Type":"ContainerStarted","Data":"c612ad7e3e8d78c4352b60c81514bdfe09343c04e8e2a535975bdfd415b32ea4"} Dec 01 08:35:42 crc kubenswrapper[4744]: I1201 08:35:42.824497 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685cbf69b9-jrc9r" event={"ID":"09924952-c000-422a-b471-e738246ae161","Type":"ContainerStarted","Data":"4bbbf5fdce710a2bedd279768fd2657a627a798cf948db321fe2a09160807ff7"} Dec 01 08:35:42 crc kubenswrapper[4744]: I1201 08:35:42.835174 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a0a609f-e643-4e6b-b0f8-428fd792fce9","Type":"ContainerStarted","Data":"3e9efe156cc4be4376971c5a296054f22c889d63d7a2cf1180f1f935b8ec5791"} Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.228599 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.276910 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.344255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-config-data\") pod \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.344756 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-fernet-keys\") pod \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.344787 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7krvk\" (UniqueName: \"kubernetes.io/projected/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-kube-api-access-7krvk\") pod \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.344817 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-scripts\") pod \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.344880 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-combined-ca-bundle\") pod \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.344990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-credential-keys\") pod \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\" (UID: \"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.363558 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" (UID: "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.371368 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" (UID: "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.376962 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-kube-api-access-7krvk" (OuterVolumeSpecName: "kube-api-access-7krvk") pod "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" (UID: "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2"). InnerVolumeSpecName "kube-api-access-7krvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.377292 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-scripts" (OuterVolumeSpecName: "scripts") pod "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" (UID: "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.446145 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-combined-ca-bundle\") pod \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.446180 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8psvl\" (UniqueName: \"kubernetes.io/projected/95085c33-3a8b-4af5-aaef-bd718dabcf6f-kube-api-access-8psvl\") pod \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.446390 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-scripts\") pod \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.446436 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95085c33-3a8b-4af5-aaef-bd718dabcf6f-logs\") pod \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.446526 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-config-data\") pod \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\" (UID: \"95085c33-3a8b-4af5-aaef-bd718dabcf6f\") " Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.447017 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.447051 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.447062 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7krvk\" (UniqueName: \"kubernetes.io/projected/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-kube-api-access-7krvk\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.447073 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.450495 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" (UID: "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.451830 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95085c33-3a8b-4af5-aaef-bd718dabcf6f-logs" (OuterVolumeSpecName: "logs") pod "95085c33-3a8b-4af5-aaef-bd718dabcf6f" (UID: "95085c33-3a8b-4af5-aaef-bd718dabcf6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.452638 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95085c33-3a8b-4af5-aaef-bd718dabcf6f-kube-api-access-8psvl" (OuterVolumeSpecName: "kube-api-access-8psvl") pod "95085c33-3a8b-4af5-aaef-bd718dabcf6f" (UID: "95085c33-3a8b-4af5-aaef-bd718dabcf6f"). InnerVolumeSpecName "kube-api-access-8psvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.453895 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-scripts" (OuterVolumeSpecName: "scripts") pod "95085c33-3a8b-4af5-aaef-bd718dabcf6f" (UID: "95085c33-3a8b-4af5-aaef-bd718dabcf6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.463522 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-config-data" (OuterVolumeSpecName: "config-data") pod "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" (UID: "cf20b7a9-6976-4f37-8cfc-f3c67d2057c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.493187 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95085c33-3a8b-4af5-aaef-bd718dabcf6f" (UID: "95085c33-3a8b-4af5-aaef-bd718dabcf6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.494554 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-config-data" (OuterVolumeSpecName: "config-data") pod "95085c33-3a8b-4af5-aaef-bd718dabcf6f" (UID: "95085c33-3a8b-4af5-aaef-bd718dabcf6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.548706 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.548736 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.548745 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95085c33-3a8b-4af5-aaef-bd718dabcf6f-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.548753 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.548763 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.548770 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95085c33-3a8b-4af5-aaef-bd718dabcf6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.548779 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8psvl\" (UniqueName: \"kubernetes.io/projected/95085c33-3a8b-4af5-aaef-bd718dabcf6f-kube-api-access-8psvl\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.825778 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-77497cdd8f-s4w2q"] Dec 01 08:35:43 crc kubenswrapper[4744]: E1201 08:35:43.826529 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95085c33-3a8b-4af5-aaef-bd718dabcf6f" containerName="placement-db-sync" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.826546 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="95085c33-3a8b-4af5-aaef-bd718dabcf6f" containerName="placement-db-sync" Dec 01 08:35:43 crc kubenswrapper[4744]: E1201 08:35:43.826578 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" containerName="keystone-bootstrap" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.826586 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" containerName="keystone-bootstrap" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.826796 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" containerName="keystone-bootstrap" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.826814 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="95085c33-3a8b-4af5-aaef-bd718dabcf6f" containerName="placement-db-sync" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.827584 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.829883 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.830155 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.856395 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77497cdd8f-s4w2q"] Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.876025 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bzxk9" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.876021 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bzxk9" event={"ID":"95085c33-3a8b-4af5-aaef-bd718dabcf6f","Type":"ContainerDied","Data":"56a818bac339ebe28970e1dc9ae87d8307a49bebc8a7f595d78f7b1653a97d4c"} Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.877463 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56a818bac339ebe28970e1dc9ae87d8307a49bebc8a7f595d78f7b1653a97d4c" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.881605 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdnlb" event={"ID":"cf20b7a9-6976-4f37-8cfc-f3c67d2057c2","Type":"ContainerDied","Data":"4b63c38ac0adbb05c96f5bf132cfb3403c3c6d2c663f7720b224ef36a3879403"} Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.881651 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b63c38ac0adbb05c96f5bf132cfb3403c3c6d2c663f7720b224ef36a3879403" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.881709 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdnlb" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.900571 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1691c44a-f2ba-43eb-9ff3-a6a25b352d78","Type":"ContainerStarted","Data":"e6cd242e2347cb88215c4090a3ce52c86d093b6f7e074e034a268a8206eb0394"} Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.915096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-685cbf69b9-jrc9r" event={"ID":"09924952-c000-422a-b471-e738246ae161","Type":"ContainerStarted","Data":"8c086f4d267953ac1238fa5610bf4903d19d7c8847b8ebd3661b1cdbffb37792"} Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.915174 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.921888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a0a609f-e643-4e6b-b0f8-428fd792fce9","Type":"ContainerStarted","Data":"2411f961caa796d67d32b9f30533c1102ca0802392405051a9b1df63f83097a2"} Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.959520 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-86964bb7b8-4bgsm"] Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.977312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981166 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-public-tls-certs\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981316 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-fernet-keys\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981374 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-config-data\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981458 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-scripts\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981485 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-credential-keys\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981549 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vprbv\" (UniqueName: \"kubernetes.io/projected/d2a839b7-c745-45e3-9984-3f403dbf6fce-kube-api-access-vprbv\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981600 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-combined-ca-bundle\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.981697 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-internal-tls-certs\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.982261 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.983056 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.983227 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4dm2j" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.983421 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.983531 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.983513854 podStartE2EDuration="3.983513854s" podCreationTimestamp="2025-12-01 08:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:43.933548452 +0000 UTC m=+1075.922606373" watchObservedRunningTime="2025-12-01 08:35:43.983513854 +0000 UTC m=+1075.972571775" Dec 01 08:35:43 crc kubenswrapper[4744]: I1201 08:35:43.983566 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.033949 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86964bb7b8-4bgsm"] Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.057485 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-685cbf69b9-jrc9r" podStartSLOduration=3.05745915 podStartE2EDuration="3.05745915s" podCreationTimestamp="2025-12-01 08:35:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:43.984924194 +0000 UTC m=+1075.973982115" watchObservedRunningTime="2025-12-01 08:35:44.05745915 +0000 UTC m=+1076.046517071" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.069877 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.069848262 podStartE2EDuration="4.069848262s" podCreationTimestamp="2025-12-01 08:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:44.023227045 +0000 UTC m=+1076.012284966" watchObservedRunningTime="2025-12-01 08:35:44.069848262 +0000 UTC m=+1076.058906183" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087611 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-internal-tls-certs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-fernet-keys\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087780 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-config-data\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087845 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-scripts\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-scripts\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087907 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-credential-keys\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vprbv\" (UniqueName: \"kubernetes.io/projected/d2a839b7-c745-45e3-9984-3f403dbf6fce-kube-api-access-vprbv\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.087997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-combined-ca-bundle\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.088046 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-logs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.088091 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-config-data\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.088125 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-public-tls-certs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.088145 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-internal-tls-certs\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.088199 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-public-tls-certs\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.088221 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-combined-ca-bundle\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.088238 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg8zf\" (UniqueName: \"kubernetes.io/projected/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-kube-api-access-sg8zf\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.097840 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-fernet-keys\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.100298 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-internal-tls-certs\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.101732 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-scripts\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.101951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-public-tls-certs\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.108226 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-combined-ca-bundle\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.108599 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-config-data\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.110821 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vprbv\" (UniqueName: \"kubernetes.io/projected/d2a839b7-c745-45e3-9984-3f403dbf6fce-kube-api-access-vprbv\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.115735 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2a839b7-c745-45e3-9984-3f403dbf6fce-credential-keys\") pod \"keystone-77497cdd8f-s4w2q\" (UID: \"d2a839b7-c745-45e3-9984-3f403dbf6fce\") " pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.158992 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.190686 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-logs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.190743 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-config-data\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.190761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-public-tls-certs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.190802 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-combined-ca-bundle\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.190823 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg8zf\" (UniqueName: \"kubernetes.io/projected/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-kube-api-access-sg8zf\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.190861 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-internal-tls-certs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.190898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-scripts\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.194717 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-logs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.196685 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-combined-ca-bundle\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.197954 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-scripts\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.198292 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-config-data\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.201167 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-public-tls-certs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.201299 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-internal-tls-certs\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.209538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg8zf\" (UniqueName: \"kubernetes.io/projected/b7d2acd7-91fd-4cc6-a9de-f7a3fb905192-kube-api-access-sg8zf\") pod \"placement-86964bb7b8-4bgsm\" (UID: \"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192\") " pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.309563 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.655942 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77497cdd8f-s4w2q"] Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.831458 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86964bb7b8-4bgsm"] Dec 01 08:35:44 crc kubenswrapper[4744]: W1201 08:35:44.843358 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d2acd7_91fd_4cc6_a9de_f7a3fb905192.slice/crio-3ca8e7c01e19e722545b4921e20b9fee77262b194cf667f9814539b0e884c6e9 WatchSource:0}: Error finding container 3ca8e7c01e19e722545b4921e20b9fee77262b194cf667f9814539b0e884c6e9: Status 404 returned error can't find the container with id 3ca8e7c01e19e722545b4921e20b9fee77262b194cf667f9814539b0e884c6e9 Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.929801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77497cdd8f-s4w2q" event={"ID":"d2a839b7-c745-45e3-9984-3f403dbf6fce","Type":"ContainerStarted","Data":"a945d1ac51db18e7f5404e83d7b31f9116a5697df6bf05a9db12ebc4a46af944"} Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.929842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77497cdd8f-s4w2q" event={"ID":"d2a839b7-c745-45e3-9984-3f403dbf6fce","Type":"ContainerStarted","Data":"d6fe63053a2b8423351ccf3df786ba1a0b957a623558e78a5942b311e2f2209e"} Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.930294 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.932721 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86964bb7b8-4bgsm" event={"ID":"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192","Type":"ContainerStarted","Data":"3ca8e7c01e19e722545b4921e20b9fee77262b194cf667f9814539b0e884c6e9"} Dec 01 08:35:44 crc kubenswrapper[4744]: I1201 08:35:44.955103 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-77497cdd8f-s4w2q" podStartSLOduration=1.955080666 podStartE2EDuration="1.955080666s" podCreationTimestamp="2025-12-01 08:35:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:44.950491896 +0000 UTC m=+1076.939549827" watchObservedRunningTime="2025-12-01 08:35:44.955080666 +0000 UTC m=+1076.944138587" Dec 01 08:35:45 crc kubenswrapper[4744]: I1201 08:35:45.944291 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86964bb7b8-4bgsm" event={"ID":"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192","Type":"ContainerStarted","Data":"686583b20def65ca70a6c8c41f413997d44b09786354f894f5e93c4f2de29ea8"} Dec 01 08:35:48 crc kubenswrapper[4744]: I1201 08:35:48.988192 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerStarted","Data":"f9cd085e1667dd05ef89c0f67fc05ed673d4ea549f9b6f9e092c6c86933bfa95"} Dec 01 08:35:48 crc kubenswrapper[4744]: I1201 08:35:48.991796 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86964bb7b8-4bgsm" event={"ID":"b7d2acd7-91fd-4cc6-a9de-f7a3fb905192","Type":"ContainerStarted","Data":"2f6b8b6c00a8ee8ada693907a513f50217908d48f2a94ec80b7e0c4cb4c7253f"} Dec 01 08:35:48 crc kubenswrapper[4744]: I1201 08:35:48.993120 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:48 crc kubenswrapper[4744]: I1201 08:35:48.993157 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.022107 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-86964bb7b8-4bgsm" podStartSLOduration=6.022081393 podStartE2EDuration="6.022081393s" podCreationTimestamp="2025-12-01 08:35:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:35:49.019325664 +0000 UTC m=+1081.008383585" watchObservedRunningTime="2025-12-01 08:35:49.022081393 +0000 UTC m=+1081.011139314" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.202262 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.270603 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-5d7w6"] Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.270879 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" podUID="9969b75a-1def-4af6-a70b-a89129668ffa" containerName="dnsmasq-dns" containerID="cri-o://74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228" gracePeriod=10 Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.707691 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.802215 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-nb\") pod \"9969b75a-1def-4af6-a70b-a89129668ffa\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.802335 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-swift-storage-0\") pod \"9969b75a-1def-4af6-a70b-a89129668ffa\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.802430 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-sb\") pod \"9969b75a-1def-4af6-a70b-a89129668ffa\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.802495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8lhc\" (UniqueName: \"kubernetes.io/projected/9969b75a-1def-4af6-a70b-a89129668ffa-kube-api-access-t8lhc\") pod \"9969b75a-1def-4af6-a70b-a89129668ffa\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.802519 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-config\") pod \"9969b75a-1def-4af6-a70b-a89129668ffa\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.802541 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-svc\") pod \"9969b75a-1def-4af6-a70b-a89129668ffa\" (UID: \"9969b75a-1def-4af6-a70b-a89129668ffa\") " Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.828760 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9969b75a-1def-4af6-a70b-a89129668ffa-kube-api-access-t8lhc" (OuterVolumeSpecName: "kube-api-access-t8lhc") pod "9969b75a-1def-4af6-a70b-a89129668ffa" (UID: "9969b75a-1def-4af6-a70b-a89129668ffa"). InnerVolumeSpecName "kube-api-access-t8lhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.857440 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9969b75a-1def-4af6-a70b-a89129668ffa" (UID: "9969b75a-1def-4af6-a70b-a89129668ffa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.863539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9969b75a-1def-4af6-a70b-a89129668ffa" (UID: "9969b75a-1def-4af6-a70b-a89129668ffa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.872853 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9969b75a-1def-4af6-a70b-a89129668ffa" (UID: "9969b75a-1def-4af6-a70b-a89129668ffa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.875813 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-config" (OuterVolumeSpecName: "config") pod "9969b75a-1def-4af6-a70b-a89129668ffa" (UID: "9969b75a-1def-4af6-a70b-a89129668ffa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.900153 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9969b75a-1def-4af6-a70b-a89129668ffa" (UID: "9969b75a-1def-4af6-a70b-a89129668ffa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.904648 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8lhc\" (UniqueName: \"kubernetes.io/projected/9969b75a-1def-4af6-a70b-a89129668ffa-kube-api-access-t8lhc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.904683 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.904692 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.904702 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.904710 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:49 crc kubenswrapper[4744]: I1201 08:35:49.904720 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9969b75a-1def-4af6-a70b-a89129668ffa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.001097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2mxc" event={"ID":"d3718521-a22d-4350-aa32-98987bfe5632","Type":"ContainerStarted","Data":"6ab49ff608f7526926bb7b38c3b848d5dea0d006ccc3a811638cf2b9c3a16ca1"} Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.002187 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-x664l" event={"ID":"ab3b54ed-348a-4314-8e92-3e977ee8f1ff","Type":"ContainerStarted","Data":"772e5a2097a6c125df66c7652c0e9b97632c6ed6ab9088063178831126e57bab"} Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.004474 4744 generic.go:334] "Generic (PLEG): container finished" podID="9969b75a-1def-4af6-a70b-a89129668ffa" containerID="74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228" exitCode=0 Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.005510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" event={"ID":"9969b75a-1def-4af6-a70b-a89129668ffa","Type":"ContainerDied","Data":"74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228"} Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.005551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" event={"ID":"9969b75a-1def-4af6-a70b-a89129668ffa","Type":"ContainerDied","Data":"8d8d0e6247b1dacd9c56c9be9abdba58b8b423643a2a934a534a270194950a5b"} Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.005574 4744 scope.go:117] "RemoveContainer" containerID="74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.005704 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-5d7w6" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.050225 4744 scope.go:117] "RemoveContainer" containerID="6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.062667 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-t2mxc" podStartSLOduration=4.041085156 podStartE2EDuration="45.06264566s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="2025-12-01 08:35:07.826472373 +0000 UTC m=+1039.815530284" lastFinishedPulling="2025-12-01 08:35:48.848032867 +0000 UTC m=+1080.837090788" observedRunningTime="2025-12-01 08:35:50.047723605 +0000 UTC m=+1082.036781526" watchObservedRunningTime="2025-12-01 08:35:50.06264566 +0000 UTC m=+1082.051703581" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.091482 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-x664l" podStartSLOduration=4.078004238 podStartE2EDuration="45.091459291s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="2025-12-01 08:35:07.829684495 +0000 UTC m=+1039.818742426" lastFinishedPulling="2025-12-01 08:35:48.843139518 +0000 UTC m=+1080.832197479" observedRunningTime="2025-12-01 08:35:50.075075744 +0000 UTC m=+1082.064133665" watchObservedRunningTime="2025-12-01 08:35:50.091459291 +0000 UTC m=+1082.080517212" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.117111 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-5d7w6"] Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.120847 4744 scope.go:117] "RemoveContainer" containerID="74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228" Dec 01 08:35:50 crc kubenswrapper[4744]: E1201 08:35:50.121323 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228\": container with ID starting with 74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228 not found: ID does not exist" containerID="74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.121359 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228"} err="failed to get container status \"74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228\": rpc error: code = NotFound desc = could not find container \"74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228\": container with ID starting with 74c5d43c8fb5074b0c7cb3cd6aeb96a893cc26c49bd2f340430c1b6e3a51e228 not found: ID does not exist" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.121376 4744 scope.go:117] "RemoveContainer" containerID="6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5" Dec 01 08:35:50 crc kubenswrapper[4744]: E1201 08:35:50.129688 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5\": container with ID starting with 6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5 not found: ID does not exist" containerID="6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.129745 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5"} err="failed to get container status \"6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5\": rpc error: code = NotFound desc = could not find container \"6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5\": container with ID starting with 6a2dcfbd080d6613108da94840e4c30d3b3894bff5fa9ab580e41ee897afc4b5 not found: ID does not exist" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.133346 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-5d7w6"] Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.298732 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9969b75a-1def-4af6-a70b-a89129668ffa" path="/var/lib/kubelet/pods/9969b75a-1def-4af6-a70b-a89129668ffa/volumes" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.596438 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.596488 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.638027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.641015 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.726601 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.726641 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.754697 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:50 crc kubenswrapper[4744]: I1201 08:35:50.773667 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:51 crc kubenswrapper[4744]: I1201 08:35:51.017764 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8mwhk" event={"ID":"ed2a24ae-749f-4faf-b077-060df14120bb","Type":"ContainerStarted","Data":"a8b3f04302e8835bc76b504b46fc1c5e954d957f2e75d1b3ece12306aae68748"} Dec 01 08:35:51 crc kubenswrapper[4744]: I1201 08:35:51.021991 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 08:35:51 crc kubenswrapper[4744]: I1201 08:35:51.022239 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:51 crc kubenswrapper[4744]: I1201 08:35:51.022359 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 08:35:51 crc kubenswrapper[4744]: I1201 08:35:51.022439 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:51 crc kubenswrapper[4744]: I1201 08:35:51.054969 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-8mwhk" podStartSLOduration=4.125378646 podStartE2EDuration="46.054944694s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="2025-12-01 08:35:07.875014585 +0000 UTC m=+1039.864072506" lastFinishedPulling="2025-12-01 08:35:49.804580633 +0000 UTC m=+1081.793638554" observedRunningTime="2025-12-01 08:35:51.038066023 +0000 UTC m=+1083.027123944" watchObservedRunningTime="2025-12-01 08:35:51.054944694 +0000 UTC m=+1083.044002615" Dec 01 08:35:51 crc kubenswrapper[4744]: I1201 08:35:51.380531 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:35:53 crc kubenswrapper[4744]: I1201 08:35:53.004741 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 08:35:53 crc kubenswrapper[4744]: I1201 08:35:53.031790 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:35:53 crc kubenswrapper[4744]: I1201 08:35:53.031819 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:35:53 crc kubenswrapper[4744]: I1201 08:35:53.031858 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:35:53 crc kubenswrapper[4744]: I1201 08:35:53.203352 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 08:35:53 crc kubenswrapper[4744]: I1201 08:35:53.276847 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:53 crc kubenswrapper[4744]: I1201 08:35:53.282864 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 08:35:55 crc kubenswrapper[4744]: I1201 08:35:55.052218 4744 generic.go:334] "Generic (PLEG): container finished" podID="ed2a24ae-749f-4faf-b077-060df14120bb" containerID="a8b3f04302e8835bc76b504b46fc1c5e954d957f2e75d1b3ece12306aae68748" exitCode=0 Dec 01 08:35:55 crc kubenswrapper[4744]: I1201 08:35:55.052290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8mwhk" event={"ID":"ed2a24ae-749f-4faf-b077-060df14120bb","Type":"ContainerDied","Data":"a8b3f04302e8835bc76b504b46fc1c5e954d957f2e75d1b3ece12306aae68748"} Dec 01 08:35:56 crc kubenswrapper[4744]: I1201 08:35:56.061890 4744 generic.go:334] "Generic (PLEG): container finished" podID="d3718521-a22d-4350-aa32-98987bfe5632" containerID="6ab49ff608f7526926bb7b38c3b848d5dea0d006ccc3a811638cf2b9c3a16ca1" exitCode=0 Dec 01 08:35:56 crc kubenswrapper[4744]: I1201 08:35:56.062016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2mxc" event={"ID":"d3718521-a22d-4350-aa32-98987bfe5632","Type":"ContainerDied","Data":"6ab49ff608f7526926bb7b38c3b848d5dea0d006ccc3a811638cf2b9c3a16ca1"} Dec 01 08:35:56 crc kubenswrapper[4744]: I1201 08:35:56.064629 4744 generic.go:334] "Generic (PLEG): container finished" podID="ab3b54ed-348a-4314-8e92-3e977ee8f1ff" containerID="772e5a2097a6c125df66c7652c0e9b97632c6ed6ab9088063178831126e57bab" exitCode=0 Dec 01 08:35:56 crc kubenswrapper[4744]: I1201 08:35:56.064707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-x664l" event={"ID":"ab3b54ed-348a-4314-8e92-3e977ee8f1ff","Type":"ContainerDied","Data":"772e5a2097a6c125df66c7652c0e9b97632c6ed6ab9088063178831126e57bab"} Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.100121 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8mwhk" event={"ID":"ed2a24ae-749f-4faf-b077-060df14120bb","Type":"ContainerDied","Data":"f308ec7a37d15fbcb510a5235947b4a8530ab5f7ec41c1a6ee39233bcaea93d2"} Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.100525 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f308ec7a37d15fbcb510a5235947b4a8530ab5f7ec41c1a6ee39233bcaea93d2" Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.146764 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.268417 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9jth\" (UniqueName: \"kubernetes.io/projected/ed2a24ae-749f-4faf-b077-060df14120bb-kube-api-access-m9jth\") pod \"ed2a24ae-749f-4faf-b077-060df14120bb\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.268465 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-combined-ca-bundle\") pod \"ed2a24ae-749f-4faf-b077-060df14120bb\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.268570 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-db-sync-config-data\") pod \"ed2a24ae-749f-4faf-b077-060df14120bb\" (UID: \"ed2a24ae-749f-4faf-b077-060df14120bb\") " Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.288717 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed2a24ae-749f-4faf-b077-060df14120bb-kube-api-access-m9jth" (OuterVolumeSpecName: "kube-api-access-m9jth") pod "ed2a24ae-749f-4faf-b077-060df14120bb" (UID: "ed2a24ae-749f-4faf-b077-060df14120bb"). InnerVolumeSpecName "kube-api-access-m9jth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.329910 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ed2a24ae-749f-4faf-b077-060df14120bb" (UID: "ed2a24ae-749f-4faf-b077-060df14120bb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.339863 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed2a24ae-749f-4faf-b077-060df14120bb" (UID: "ed2a24ae-749f-4faf-b077-060df14120bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.370891 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9jth\" (UniqueName: \"kubernetes.io/projected/ed2a24ae-749f-4faf-b077-060df14120bb-kube-api-access-m9jth\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.370941 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:57 crc kubenswrapper[4744]: I1201 08:35:57.370951 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ed2a24ae-749f-4faf-b077-060df14120bb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.107120 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8mwhk" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.371056 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7b8696b95c-fb5g6"] Dec 01 08:35:58 crc kubenswrapper[4744]: E1201 08:35:58.371425 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9969b75a-1def-4af6-a70b-a89129668ffa" containerName="dnsmasq-dns" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.371438 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9969b75a-1def-4af6-a70b-a89129668ffa" containerName="dnsmasq-dns" Dec 01 08:35:58 crc kubenswrapper[4744]: E1201 08:35:58.371453 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2a24ae-749f-4faf-b077-060df14120bb" containerName="barbican-db-sync" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.371460 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2a24ae-749f-4faf-b077-060df14120bb" containerName="barbican-db-sync" Dec 01 08:35:58 crc kubenswrapper[4744]: E1201 08:35:58.371478 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9969b75a-1def-4af6-a70b-a89129668ffa" containerName="init" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.371484 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9969b75a-1def-4af6-a70b-a89129668ffa" containerName="init" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.371637 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9969b75a-1def-4af6-a70b-a89129668ffa" containerName="dnsmasq-dns" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.371650 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2a24ae-749f-4faf-b077-060df14120bb" containerName="barbican-db-sync" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.376766 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.380915 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mgtfz" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.381217 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.384119 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.387592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-config-data\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.387657 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a23750f1-c129-46b5-9c30-c6fa8ee54320-logs\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.387697 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhrqg\" (UniqueName: \"kubernetes.io/projected/a23750f1-c129-46b5-9c30-c6fa8ee54320-kube-api-access-bhrqg\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.387719 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-config-data-custom\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.387743 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-combined-ca-bundle\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.391486 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b8696b95c-fb5g6"] Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.490035 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-config-data\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.490480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a23750f1-c129-46b5-9c30-c6fa8ee54320-logs\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.490530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhrqg\" (UniqueName: \"kubernetes.io/projected/a23750f1-c129-46b5-9c30-c6fa8ee54320-kube-api-access-bhrqg\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.490559 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-config-data-custom\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.490613 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-combined-ca-bundle\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.492342 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a23750f1-c129-46b5-9c30-c6fa8ee54320-logs\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.495789 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-config-data\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.499056 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-combined-ca-bundle\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.505461 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23750f1-c129-46b5-9c30-c6fa8ee54320-config-data-custom\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.546383 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhrqg\" (UniqueName: \"kubernetes.io/projected/a23750f1-c129-46b5-9c30-c6fa8ee54320-kube-api-access-bhrqg\") pod \"barbican-worker-7b8696b95c-fb5g6\" (UID: \"a23750f1-c129-46b5-9c30-c6fa8ee54320\") " pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.554507 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-99tfx"] Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.559391 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.606252 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7f8b9684d-5kjvj"] Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.607741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.611562 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.628103 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-99tfx"] Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.642305 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f8b9684d-5kjvj"] Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.668190 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-556674cd78-bw7k7"] Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.670151 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.673219 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.689074 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-556674cd78-bw7k7"] Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.695141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-svc\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.695344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.695539 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.695636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.695733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-config\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.695884 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5zmz\" (UniqueName: \"kubernetes.io/projected/8a458601-7873-4c35-9da5-412411380de8-kube-api-access-m5zmz\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.699136 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b8696b95c-fb5g6" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797382 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-config\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797464 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm5kr\" (UniqueName: \"kubernetes.io/projected/02654eb7-3943-40d1-9145-04bcd38b8344-kube-api-access-zm5kr\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02654eb7-3943-40d1-9145-04bcd38b8344-logs\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797538 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data-custom\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797576 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5zmz\" (UniqueName: \"kubernetes.io/projected/8a458601-7873-4c35-9da5-412411380de8-kube-api-access-m5zmz\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797629 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1473d13-7648-4ae3-8919-0bd62c6893e8-logs\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797683 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-config-data\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797706 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-combined-ca-bundle\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797732 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-config-data-custom\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797759 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-svc\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797786 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrbcz\" (UniqueName: \"kubernetes.io/projected/f1473d13-7648-4ae3-8919-0bd62c6893e8-kube-api-access-mrbcz\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797841 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-combined-ca-bundle\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.797917 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.798292 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-config\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.798997 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.799070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-svc\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.799542 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.800469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.821814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5zmz\" (UniqueName: \"kubernetes.io/projected/8a458601-7873-4c35-9da5-412411380de8-kube-api-access-m5zmz\") pod \"dnsmasq-dns-688c87cc99-99tfx\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899510 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1473d13-7648-4ae3-8919-0bd62c6893e8-logs\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-config-data\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899660 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-combined-ca-bundle\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-config-data-custom\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrbcz\" (UniqueName: \"kubernetes.io/projected/f1473d13-7648-4ae3-8919-0bd62c6893e8-kube-api-access-mrbcz\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899732 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-combined-ca-bundle\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899760 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899798 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm5kr\" (UniqueName: \"kubernetes.io/projected/02654eb7-3943-40d1-9145-04bcd38b8344-kube-api-access-zm5kr\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899840 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02654eb7-3943-40d1-9145-04bcd38b8344-logs\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.899859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data-custom\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.900100 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1473d13-7648-4ae3-8919-0bd62c6893e8-logs\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.901264 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02654eb7-3943-40d1-9145-04bcd38b8344-logs\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.904923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-combined-ca-bundle\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.905242 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-combined-ca-bundle\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.906431 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.907389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-config-data-custom\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.914876 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02654eb7-3943-40d1-9145-04bcd38b8344-config-data\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.914927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data-custom\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.917754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrbcz\" (UniqueName: \"kubernetes.io/projected/f1473d13-7648-4ae3-8919-0bd62c6893e8-kube-api-access-mrbcz\") pod \"barbican-api-556674cd78-bw7k7\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.920773 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm5kr\" (UniqueName: \"kubernetes.io/projected/02654eb7-3943-40d1-9145-04bcd38b8344-kube-api-access-zm5kr\") pod \"barbican-keystone-listener-7f8b9684d-5kjvj\" (UID: \"02654eb7-3943-40d1-9145-04bcd38b8344\") " pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.942993 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.963870 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" Dec 01 08:35:58 crc kubenswrapper[4744]: I1201 08:35:58.991982 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.114276 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-x664l" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.117308 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.149016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-t2mxc" event={"ID":"d3718521-a22d-4350-aa32-98987bfe5632","Type":"ContainerDied","Data":"79de33a4f6fc4753edcae6ad455721d3e9d5338c1321747dc34489f7bac58534"} Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.149070 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79de33a4f6fc4753edcae6ad455721d3e9d5338c1321747dc34489f7bac58534" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.149157 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-t2mxc" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.156365 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-x664l" event={"ID":"ab3b54ed-348a-4314-8e92-3e977ee8f1ff","Type":"ContainerDied","Data":"defe48eadc0573a912670898c2992ef4c6762f0c03f98124303c4624e374650c"} Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.156401 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="defe48eadc0573a912670898c2992ef4c6762f0c03f98124303c4624e374650c" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.156475 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-x664l" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.203868 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3718521-a22d-4350-aa32-98987bfe5632-etc-machine-id\") pod \"d3718521-a22d-4350-aa32-98987bfe5632\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.203968 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-config-data\") pod \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204004 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-scripts\") pod \"d3718521-a22d-4350-aa32-98987bfe5632\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204049 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3718521-a22d-4350-aa32-98987bfe5632-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d3718521-a22d-4350-aa32-98987bfe5632" (UID: "d3718521-a22d-4350-aa32-98987bfe5632"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-combined-ca-bundle\") pod \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-db-sync-config-data\") pod \"d3718521-a22d-4350-aa32-98987bfe5632\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-combined-ca-bundle\") pod \"d3718521-a22d-4350-aa32-98987bfe5632\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204144 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv46w\" (UniqueName: \"kubernetes.io/projected/d3718521-a22d-4350-aa32-98987bfe5632-kube-api-access-wv46w\") pod \"d3718521-a22d-4350-aa32-98987bfe5632\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204165 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nj56\" (UniqueName: \"kubernetes.io/projected/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-kube-api-access-4nj56\") pod \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\" (UID: \"ab3b54ed-348a-4314-8e92-3e977ee8f1ff\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-config-data\") pod \"d3718521-a22d-4350-aa32-98987bfe5632\" (UID: \"d3718521-a22d-4350-aa32-98987bfe5632\") " Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.204607 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3718521-a22d-4350-aa32-98987bfe5632-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.209797 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3718521-a22d-4350-aa32-98987bfe5632-kube-api-access-wv46w" (OuterVolumeSpecName: "kube-api-access-wv46w") pod "d3718521-a22d-4350-aa32-98987bfe5632" (UID: "d3718521-a22d-4350-aa32-98987bfe5632"). InnerVolumeSpecName "kube-api-access-wv46w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.210354 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-kube-api-access-4nj56" (OuterVolumeSpecName: "kube-api-access-4nj56") pod "ab3b54ed-348a-4314-8e92-3e977ee8f1ff" (UID: "ab3b54ed-348a-4314-8e92-3e977ee8f1ff"). InnerVolumeSpecName "kube-api-access-4nj56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.209498 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-scripts" (OuterVolumeSpecName: "scripts") pod "d3718521-a22d-4350-aa32-98987bfe5632" (UID: "d3718521-a22d-4350-aa32-98987bfe5632"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.212179 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d3718521-a22d-4350-aa32-98987bfe5632" (UID: "d3718521-a22d-4350-aa32-98987bfe5632"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.243746 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3718521-a22d-4350-aa32-98987bfe5632" (UID: "d3718521-a22d-4350-aa32-98987bfe5632"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.268389 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab3b54ed-348a-4314-8e92-3e977ee8f1ff" (UID: "ab3b54ed-348a-4314-8e92-3e977ee8f1ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.269513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-config-data" (OuterVolumeSpecName: "config-data") pod "d3718521-a22d-4350-aa32-98987bfe5632" (UID: "d3718521-a22d-4350-aa32-98987bfe5632"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.305681 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.305711 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.305721 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.305729 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.305738 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv46w\" (UniqueName: \"kubernetes.io/projected/d3718521-a22d-4350-aa32-98987bfe5632-kube-api-access-wv46w\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.305747 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nj56\" (UniqueName: \"kubernetes.io/projected/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-kube-api-access-4nj56\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.305755 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3718521-a22d-4350-aa32-98987bfe5632-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.309736 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-config-data" (OuterVolumeSpecName: "config-data") pod "ab3b54ed-348a-4314-8e92-3e977ee8f1ff" (UID: "ab3b54ed-348a-4314-8e92-3e977ee8f1ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.407806 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3b54ed-348a-4314-8e92-3e977ee8f1ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.742543 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-99tfx"] Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.837204 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b8696b95c-fb5g6"] Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.922761 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f8b9684d-5kjvj"] Dec 01 08:35:59 crc kubenswrapper[4744]: W1201 08:35:59.924100 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02654eb7_3943_40d1_9145_04bcd38b8344.slice/crio-169712ed3832f3d210c6d434ac94374e890ebe0f57b2854b473f7bc87fe88bf9 WatchSource:0}: Error finding container 169712ed3832f3d210c6d434ac94374e890ebe0f57b2854b473f7bc87fe88bf9: Status 404 returned error can't find the container with id 169712ed3832f3d210c6d434ac94374e890ebe0f57b2854b473f7bc87fe88bf9 Dec 01 08:35:59 crc kubenswrapper[4744]: W1201 08:35:59.936491 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1473d13_7648_4ae3_8919_0bd62c6893e8.slice/crio-ca16c627673803853a3a62edc46b7813faa120011f502dc6d4e56093a4358a78 WatchSource:0}: Error finding container ca16c627673803853a3a62edc46b7813faa120011f502dc6d4e56093a4358a78: Status 404 returned error can't find the container with id ca16c627673803853a3a62edc46b7813faa120011f502dc6d4e56093a4358a78 Dec 01 08:35:59 crc kubenswrapper[4744]: I1201 08:35:59.937688 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-556674cd78-bw7k7"] Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.204617 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerStarted","Data":"f24d512de5bbcc73f434099c014a8e7b58530d159816f86bedd633c4372052a4"} Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.205380 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-central-agent" containerID="cri-o://179e72728e42caf2adc1f8f305005ce810fc06276b7326ef6ea701ce3d37c130" gracePeriod=30 Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.205743 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.206119 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="proxy-httpd" containerID="cri-o://f24d512de5bbcc73f434099c014a8e7b58530d159816f86bedd633c4372052a4" gracePeriod=30 Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.206187 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="sg-core" containerID="cri-o://f9cd085e1667dd05ef89c0f67fc05ed673d4ea549f9b6f9e092c6c86933bfa95" gracePeriod=30 Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.206228 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-notification-agent" containerID="cri-o://41101381762b71bd360b63ab1e8962d90cd86d9fc2e3a55615047f712ab5c16b" gracePeriod=30 Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.220343 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" event={"ID":"02654eb7-3943-40d1-9145-04bcd38b8344","Type":"ContainerStarted","Data":"169712ed3832f3d210c6d434ac94374e890ebe0f57b2854b473f7bc87fe88bf9"} Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.241206 4744 generic.go:334] "Generic (PLEG): container finished" podID="8a458601-7873-4c35-9da5-412411380de8" containerID="9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6" exitCode=0 Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.241511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" event={"ID":"8a458601-7873-4c35-9da5-412411380de8","Type":"ContainerDied","Data":"9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6"} Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.241607 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" event={"ID":"8a458601-7873-4c35-9da5-412411380de8","Type":"ContainerStarted","Data":"1089b5a37450a3dfa2db6c37285b706b0ac85ac10541b073dffc28a0e5242e01"} Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.244071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b8696b95c-fb5g6" event={"ID":"a23750f1-c129-46b5-9c30-c6fa8ee54320","Type":"ContainerStarted","Data":"a8c071564d034731d84bb363879ca0155a11c68c862f946adc07ff1c466ad46c"} Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.246198 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556674cd78-bw7k7" event={"ID":"f1473d13-7648-4ae3-8919-0bd62c6893e8","Type":"ContainerStarted","Data":"59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65"} Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.246240 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556674cd78-bw7k7" event={"ID":"f1473d13-7648-4ae3-8919-0bd62c6893e8","Type":"ContainerStarted","Data":"ca16c627673803853a3a62edc46b7813faa120011f502dc6d4e56093a4358a78"} Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.266941 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.81925924 podStartE2EDuration="55.266912908s" podCreationTimestamp="2025-12-01 08:35:05 +0000 UTC" firstStartedPulling="2025-12-01 08:35:07.831251449 +0000 UTC m=+1039.820309370" lastFinishedPulling="2025-12-01 08:35:59.278905117 +0000 UTC m=+1091.267963038" observedRunningTime="2025-12-01 08:36:00.249660357 +0000 UTC m=+1092.238718278" watchObservedRunningTime="2025-12-01 08:36:00.266912908 +0000 UTC m=+1092.255970829" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.477909 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:00 crc kubenswrapper[4744]: E1201 08:36:00.478602 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3718521-a22d-4350-aa32-98987bfe5632" containerName="cinder-db-sync" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.478694 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3718521-a22d-4350-aa32-98987bfe5632" containerName="cinder-db-sync" Dec 01 08:36:00 crc kubenswrapper[4744]: E1201 08:36:00.478772 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3b54ed-348a-4314-8e92-3e977ee8f1ff" containerName="heat-db-sync" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.478820 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3b54ed-348a-4314-8e92-3e977ee8f1ff" containerName="heat-db-sync" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.479037 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3718521-a22d-4350-aa32-98987bfe5632" containerName="cinder-db-sync" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.479106 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3b54ed-348a-4314-8e92-3e977ee8f1ff" containerName="heat-db-sync" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.480455 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.485529 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-sln7j" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.485835 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.485980 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.486121 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.494593 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.527825 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-99tfx"] Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.575352 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-5hrgd"] Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.581389 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.599239 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-5hrgd"] Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.647952 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.647995 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.648045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.648072 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.648128 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr22t\" (UniqueName: \"kubernetes.io/projected/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-kube-api-access-pr22t\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.648168 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-scripts\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smwl9\" (UniqueName: \"kubernetes.io/projected/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-kube-api-access-smwl9\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750525 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750557 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750577 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750645 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750752 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750781 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750822 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-config\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750864 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr22t\" (UniqueName: \"kubernetes.io/projected/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-kube-api-access-pr22t\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.750957 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-scripts\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.755552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.756896 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-scripts\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.757322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.757359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.777628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.785981 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr22t\" (UniqueName: \"kubernetes.io/projected/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-kube-api-access-pr22t\") pod \"cinder-scheduler-0\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.786223 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.789481 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.792553 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.834747 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.852341 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.852583 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.852690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.852771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-config\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.852850 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.852955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smwl9\" (UniqueName: \"kubernetes.io/projected/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-kube-api-access-smwl9\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.853176 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.853736 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-config\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.853764 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.854060 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.854296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.879094 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smwl9\" (UniqueName: \"kubernetes.io/projected/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-kube-api-access-smwl9\") pod \"dnsmasq-dns-6bb4fc677f-5hrgd\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.895140 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.925067 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.959539 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data-custom\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.959677 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-scripts\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.959738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.959767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.959864 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnd5c\" (UniqueName: \"kubernetes.io/projected/c61efc8b-a454-412f-bc07-2a734a8bf489-kube-api-access-xnd5c\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.959887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c61efc8b-a454-412f-bc07-2a734a8bf489-logs\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:00 crc kubenswrapper[4744]: I1201 08:36:00.959951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c61efc8b-a454-412f-bc07-2a734a8bf489-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.061144 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-scripts\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.061223 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.061256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.061298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnd5c\" (UniqueName: \"kubernetes.io/projected/c61efc8b-a454-412f-bc07-2a734a8bf489-kube-api-access-xnd5c\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.061316 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c61efc8b-a454-412f-bc07-2a734a8bf489-logs\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.061356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c61efc8b-a454-412f-bc07-2a734a8bf489-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.061397 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data-custom\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.066707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c61efc8b-a454-412f-bc07-2a734a8bf489-logs\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.067015 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c61efc8b-a454-412f-bc07-2a734a8bf489-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.075108 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.075494 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-scripts\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.076184 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.076632 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data-custom\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.083774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnd5c\" (UniqueName: \"kubernetes.io/projected/c61efc8b-a454-412f-bc07-2a734a8bf489-kube-api-access-xnd5c\") pod \"cinder-api-0\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.279523 4744 generic.go:334] "Generic (PLEG): container finished" podID="96a23a18-05c4-409a-950d-0743db8fc652" containerID="f24d512de5bbcc73f434099c014a8e7b58530d159816f86bedd633c4372052a4" exitCode=0 Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.279919 4744 generic.go:334] "Generic (PLEG): container finished" podID="96a23a18-05c4-409a-950d-0743db8fc652" containerID="f9cd085e1667dd05ef89c0f67fc05ed673d4ea549f9b6f9e092c6c86933bfa95" exitCode=2 Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.279935 4744 generic.go:334] "Generic (PLEG): container finished" podID="96a23a18-05c4-409a-950d-0743db8fc652" containerID="179e72728e42caf2adc1f8f305005ce810fc06276b7326ef6ea701ce3d37c130" exitCode=0 Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.280002 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerDied","Data":"f24d512de5bbcc73f434099c014a8e7b58530d159816f86bedd633c4372052a4"} Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.280035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerDied","Data":"f9cd085e1667dd05ef89c0f67fc05ed673d4ea549f9b6f9e092c6c86933bfa95"} Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.280056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerDied","Data":"179e72728e42caf2adc1f8f305005ce810fc06276b7326ef6ea701ce3d37c130"} Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.283386 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556674cd78-bw7k7" event={"ID":"f1473d13-7648-4ae3-8919-0bd62c6893e8","Type":"ContainerStarted","Data":"5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3"} Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.283951 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.284141 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.284250 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.288854 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" event={"ID":"8a458601-7873-4c35-9da5-412411380de8","Type":"ContainerStarted","Data":"0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7"} Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.289022 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" podUID="8a458601-7873-4c35-9da5-412411380de8" containerName="dnsmasq-dns" containerID="cri-o://0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7" gracePeriod=10 Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.289225 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.311321 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-556674cd78-bw7k7" podStartSLOduration=3.311303545 podStartE2EDuration="3.311303545s" podCreationTimestamp="2025-12-01 08:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:01.308177026 +0000 UTC m=+1093.297234947" watchObservedRunningTime="2025-12-01 08:36:01.311303545 +0000 UTC m=+1093.300361466" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.377729 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" podStartSLOduration=3.377709025 podStartE2EDuration="3.377709025s" podCreationTimestamp="2025-12-01 08:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:01.33502729 +0000 UTC m=+1093.324085211" watchObservedRunningTime="2025-12-01 08:36:01.377709025 +0000 UTC m=+1093.366766956" Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.393067 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:01 crc kubenswrapper[4744]: W1201 08:36:01.406775 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa35ddd_0dca_46e2_8f4f_49060a63d45c.slice/crio-7be580839ea42e8ed7878d1e6b6936ce0ddbc1f91772ad1cd6f06d7ac17bd19d WatchSource:0}: Error finding container 7be580839ea42e8ed7878d1e6b6936ce0ddbc1f91772ad1cd6f06d7ac17bd19d: Status 404 returned error can't find the container with id 7be580839ea42e8ed7878d1e6b6936ce0ddbc1f91772ad1cd6f06d7ac17bd19d Dec 01 08:36:01 crc kubenswrapper[4744]: W1201 08:36:01.475130 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87a3a2b6_e743_46dd_bcdb_a6f598ac8c64.slice/crio-dab3eedac35259b7bb90eed5949e5f4ffc756690766947672bec32b12d368831 WatchSource:0}: Error finding container dab3eedac35259b7bb90eed5949e5f4ffc756690766947672bec32b12d368831: Status 404 returned error can't find the container with id dab3eedac35259b7bb90eed5949e5f4ffc756690766947672bec32b12d368831 Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.475857 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-5hrgd"] Dec 01 08:36:01 crc kubenswrapper[4744]: I1201 08:36:01.815865 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.174627 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.294954 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-nb\") pod \"8a458601-7873-4c35-9da5-412411380de8\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.295023 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-svc\") pod \"8a458601-7873-4c35-9da5-412411380de8\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.295052 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-swift-storage-0\") pod \"8a458601-7873-4c35-9da5-412411380de8\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.295124 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-sb\") pod \"8a458601-7873-4c35-9da5-412411380de8\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.295219 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-config\") pod \"8a458601-7873-4c35-9da5-412411380de8\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.295319 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5zmz\" (UniqueName: \"kubernetes.io/projected/8a458601-7873-4c35-9da5-412411380de8-kube-api-access-m5zmz\") pod \"8a458601-7873-4c35-9da5-412411380de8\" (UID: \"8a458601-7873-4c35-9da5-412411380de8\") " Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.299815 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a458601-7873-4c35-9da5-412411380de8-kube-api-access-m5zmz" (OuterVolumeSpecName: "kube-api-access-m5zmz") pod "8a458601-7873-4c35-9da5-412411380de8" (UID: "8a458601-7873-4c35-9da5-412411380de8"). InnerVolumeSpecName "kube-api-access-m5zmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.313332 4744 generic.go:334] "Generic (PLEG): container finished" podID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerID="3d9f0daedf8f3b1b963778314cf5092b15f5b402e3e059242fa9404b244dc83b" exitCode=0 Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.313470 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" event={"ID":"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64","Type":"ContainerDied","Data":"3d9f0daedf8f3b1b963778314cf5092b15f5b402e3e059242fa9404b244dc83b"} Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.313538 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" event={"ID":"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64","Type":"ContainerStarted","Data":"dab3eedac35259b7bb90eed5949e5f4ffc756690766947672bec32b12d368831"} Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.315653 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c61efc8b-a454-412f-bc07-2a734a8bf489","Type":"ContainerStarted","Data":"0a261a421293bfdefef719a28b5736ceed1f6ab0216584c73cf416d296df3fbf"} Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.317444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0aa35ddd-0dca-46e2-8f4f-49060a63d45c","Type":"ContainerStarted","Data":"7be580839ea42e8ed7878d1e6b6936ce0ddbc1f91772ad1cd6f06d7ac17bd19d"} Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.320050 4744 generic.go:334] "Generic (PLEG): container finished" podID="8a458601-7873-4c35-9da5-412411380de8" containerID="0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7" exitCode=0 Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.320123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" event={"ID":"8a458601-7873-4c35-9da5-412411380de8","Type":"ContainerDied","Data":"0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7"} Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.320146 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.320175 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-99tfx" event={"ID":"8a458601-7873-4c35-9da5-412411380de8","Type":"ContainerDied","Data":"1089b5a37450a3dfa2db6c37285b706b0ac85ac10541b073dffc28a0e5242e01"} Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.320198 4744 scope.go:117] "RemoveContainer" containerID="0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.356391 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8a458601-7873-4c35-9da5-412411380de8" (UID: "8a458601-7873-4c35-9da5-412411380de8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.358045 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-config" (OuterVolumeSpecName: "config") pod "8a458601-7873-4c35-9da5-412411380de8" (UID: "8a458601-7873-4c35-9da5-412411380de8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.366179 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8a458601-7873-4c35-9da5-412411380de8" (UID: "8a458601-7873-4c35-9da5-412411380de8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.366599 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8a458601-7873-4c35-9da5-412411380de8" (UID: "8a458601-7873-4c35-9da5-412411380de8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.373441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8a458601-7873-4c35-9da5-412411380de8" (UID: "8a458601-7873-4c35-9da5-412411380de8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.398267 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.398312 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.398328 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.398339 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.398351 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a458601-7873-4c35-9da5-412411380de8-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.398363 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5zmz\" (UniqueName: \"kubernetes.io/projected/8a458601-7873-4c35-9da5-412411380de8-kube-api-access-m5zmz\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.489746 4744 scope.go:117] "RemoveContainer" containerID="9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.525907 4744 scope.go:117] "RemoveContainer" containerID="0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7" Dec 01 08:36:02 crc kubenswrapper[4744]: E1201 08:36:02.526348 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7\": container with ID starting with 0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7 not found: ID does not exist" containerID="0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.526454 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7"} err="failed to get container status \"0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7\": rpc error: code = NotFound desc = could not find container \"0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7\": container with ID starting with 0d1fb4543f9ee50ea6852cf9bb42f41208fd36a7177a93428197807185842cf7 not found: ID does not exist" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.526535 4744 scope.go:117] "RemoveContainer" containerID="9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6" Dec 01 08:36:02 crc kubenswrapper[4744]: E1201 08:36:02.526845 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6\": container with ID starting with 9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6 not found: ID does not exist" containerID="9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.526889 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6"} err="failed to get container status \"9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6\": rpc error: code = NotFound desc = could not find container \"9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6\": container with ID starting with 9002c01c5122f4cea6544137b3c373b94de5c39da2ec69c3a30970e9117bd2c6 not found: ID does not exist" Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.726491 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-99tfx"] Dec 01 08:36:02 crc kubenswrapper[4744]: I1201 08:36:02.735272 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-99tfx"] Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.338280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" event={"ID":"02654eb7-3943-40d1-9145-04bcd38b8344","Type":"ContainerStarted","Data":"8ac6054f7199866b445d71cf146e4f03bd88aaffe1f3aa327e0020a90e231bf7"} Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.338710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" event={"ID":"02654eb7-3943-40d1-9145-04bcd38b8344","Type":"ContainerStarted","Data":"d21177e26b7ca74cb1bfafc3fbd20c1483edfc508ee053dc1a91fccd8db53c4f"} Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.340643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b8696b95c-fb5g6" event={"ID":"a23750f1-c129-46b5-9c30-c6fa8ee54320","Type":"ContainerStarted","Data":"86765bcac00c8846b81ba4528b7602847abb25ccb647ed74d9bce6c39ebcd7f8"} Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.340663 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b8696b95c-fb5g6" event={"ID":"a23750f1-c129-46b5-9c30-c6fa8ee54320","Type":"ContainerStarted","Data":"310706d5f5e0790f508045ce397389034e62b6a0c7877a34b7a6d6edf14c3137"} Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.359260 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7f8b9684d-5kjvj" podStartSLOduration=2.756159797 podStartE2EDuration="5.359243033s" podCreationTimestamp="2025-12-01 08:35:58 +0000 UTC" firstStartedPulling="2025-12-01 08:35:59.926832655 +0000 UTC m=+1091.915890576" lastFinishedPulling="2025-12-01 08:36:02.529915891 +0000 UTC m=+1094.518973812" observedRunningTime="2025-12-01 08:36:03.355577859 +0000 UTC m=+1095.344635780" watchObservedRunningTime="2025-12-01 08:36:03.359243033 +0000 UTC m=+1095.348300954" Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.364360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" event={"ID":"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64","Type":"ContainerStarted","Data":"77e44ebd4d9ce670b2487d0b6fa54296c4d22e5d154f182a92bf21c1af24f22e"} Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.364471 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.366438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c61efc8b-a454-412f-bc07-2a734a8bf489","Type":"ContainerStarted","Data":"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327"} Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.375732 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7b8696b95c-fb5g6" podStartSLOduration=2.723191179 podStartE2EDuration="5.375711042s" podCreationTimestamp="2025-12-01 08:35:58 +0000 UTC" firstStartedPulling="2025-12-01 08:35:59.837356958 +0000 UTC m=+1091.826414869" lastFinishedPulling="2025-12-01 08:36:02.489876811 +0000 UTC m=+1094.478934732" observedRunningTime="2025-12-01 08:36:03.375345122 +0000 UTC m=+1095.364403043" watchObservedRunningTime="2025-12-01 08:36:03.375711042 +0000 UTC m=+1095.364768963" Dec 01 08:36:03 crc kubenswrapper[4744]: I1201 08:36:03.412250 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" podStartSLOduration=3.412230312 podStartE2EDuration="3.412230312s" podCreationTimestamp="2025-12-01 08:36:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:03.395289339 +0000 UTC m=+1095.384347270" watchObservedRunningTime="2025-12-01 08:36:03.412230312 +0000 UTC m=+1095.401288233" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.297738 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a458601-7873-4c35-9da5-412411380de8" path="/var/lib/kubelet/pods/8a458601-7873-4c35-9da5-412411380de8/volumes" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.378244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c61efc8b-a454-412f-bc07-2a734a8bf489","Type":"ContainerStarted","Data":"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e"} Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.378324 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.381797 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0aa35ddd-0dca-46e2-8f4f-49060a63d45c","Type":"ContainerStarted","Data":"d55982341b7a8f7f0bde85b82967b57831a2b833e37c1e8b12d618c74b7e4c27"} Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.385121 4744 generic.go:334] "Generic (PLEG): container finished" podID="96a23a18-05c4-409a-950d-0743db8fc652" containerID="41101381762b71bd360b63ab1e8962d90cd86d9fc2e3a55615047f712ab5c16b" exitCode=0 Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.386427 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerDied","Data":"41101381762b71bd360b63ab1e8962d90cd86d9fc2e3a55615047f712ab5c16b"} Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.412360 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.412336527 podStartE2EDuration="4.412336527s" podCreationTimestamp="2025-12-01 08:36:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:04.400052628 +0000 UTC m=+1096.389110549" watchObservedRunningTime="2025-12-01 08:36:04.412336527 +0000 UTC m=+1096.401394448" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.777662 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.953476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-config-data\") pod \"96a23a18-05c4-409a-950d-0743db8fc652\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.953553 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-combined-ca-bundle\") pod \"96a23a18-05c4-409a-950d-0743db8fc652\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.953589 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-sg-core-conf-yaml\") pod \"96a23a18-05c4-409a-950d-0743db8fc652\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.953705 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-scripts\") pod \"96a23a18-05c4-409a-950d-0743db8fc652\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.953763 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-run-httpd\") pod \"96a23a18-05c4-409a-950d-0743db8fc652\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.953786 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgrq9\" (UniqueName: \"kubernetes.io/projected/96a23a18-05c4-409a-950d-0743db8fc652-kube-api-access-wgrq9\") pod \"96a23a18-05c4-409a-950d-0743db8fc652\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.953838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-log-httpd\") pod \"96a23a18-05c4-409a-950d-0743db8fc652\" (UID: \"96a23a18-05c4-409a-950d-0743db8fc652\") " Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.954478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "96a23a18-05c4-409a-950d-0743db8fc652" (UID: "96a23a18-05c4-409a-950d-0743db8fc652"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.954612 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "96a23a18-05c4-409a-950d-0743db8fc652" (UID: "96a23a18-05c4-409a-950d-0743db8fc652"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.958935 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a23a18-05c4-409a-950d-0743db8fc652-kube-api-access-wgrq9" (OuterVolumeSpecName: "kube-api-access-wgrq9") pod "96a23a18-05c4-409a-950d-0743db8fc652" (UID: "96a23a18-05c4-409a-950d-0743db8fc652"). InnerVolumeSpecName "kube-api-access-wgrq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.962546 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-scripts" (OuterVolumeSpecName: "scripts") pod "96a23a18-05c4-409a-950d-0743db8fc652" (UID: "96a23a18-05c4-409a-950d-0743db8fc652"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:04 crc kubenswrapper[4744]: I1201 08:36:04.990300 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "96a23a18-05c4-409a-950d-0743db8fc652" (UID: "96a23a18-05c4-409a-950d-0743db8fc652"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.055893 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.055933 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.055950 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.055961 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96a23a18-05c4-409a-950d-0743db8fc652-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.055972 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgrq9\" (UniqueName: \"kubernetes.io/projected/96a23a18-05c4-409a-950d-0743db8fc652-kube-api-access-wgrq9\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.070249 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96a23a18-05c4-409a-950d-0743db8fc652" (UID: "96a23a18-05c4-409a-950d-0743db8fc652"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.088622 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-config-data" (OuterVolumeSpecName: "config-data") pod "96a23a18-05c4-409a-950d-0743db8fc652" (UID: "96a23a18-05c4-409a-950d-0743db8fc652"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.122875 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.157589 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.157628 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96a23a18-05c4-409a-950d-0743db8fc652-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.395696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0aa35ddd-0dca-46e2-8f4f-49060a63d45c","Type":"ContainerStarted","Data":"c8b83fd4f7450a8f5a042a0eeabaf0c5af734c24853a17d7669c43424340ed4c"} Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.398994 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.414500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96a23a18-05c4-409a-950d-0743db8fc652","Type":"ContainerDied","Data":"669a774253f2ff55d515414d17783a1443beff590a5ab59ad4798a76bbec3fc6"} Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.414562 4744 scope.go:117] "RemoveContainer" containerID="f24d512de5bbcc73f434099c014a8e7b58530d159816f86bedd633c4372052a4" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.458112 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.951782405 podStartE2EDuration="5.458088042s" podCreationTimestamp="2025-12-01 08:36:00 +0000 UTC" firstStartedPulling="2025-12-01 08:36:01.411390264 +0000 UTC m=+1093.400448185" lastFinishedPulling="2025-12-01 08:36:02.917695901 +0000 UTC m=+1094.906753822" observedRunningTime="2025-12-01 08:36:05.451936817 +0000 UTC m=+1097.440994748" watchObservedRunningTime="2025-12-01 08:36:05.458088042 +0000 UTC m=+1097.447145963" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.488667 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.490173 4744 scope.go:117] "RemoveContainer" containerID="f9cd085e1667dd05ef89c0f67fc05ed673d4ea549f9b6f9e092c6c86933bfa95" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.502324 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.515365 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:05 crc kubenswrapper[4744]: E1201 08:36:05.515743 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a458601-7873-4c35-9da5-412411380de8" containerName="init" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.515761 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a458601-7873-4c35-9da5-412411380de8" containerName="init" Dec 01 08:36:05 crc kubenswrapper[4744]: E1201 08:36:05.515780 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a458601-7873-4c35-9da5-412411380de8" containerName="dnsmasq-dns" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.515786 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a458601-7873-4c35-9da5-412411380de8" containerName="dnsmasq-dns" Dec 01 08:36:05 crc kubenswrapper[4744]: E1201 08:36:05.515809 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-central-agent" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.515815 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-central-agent" Dec 01 08:36:05 crc kubenswrapper[4744]: E1201 08:36:05.515824 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-notification-agent" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.515830 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-notification-agent" Dec 01 08:36:05 crc kubenswrapper[4744]: E1201 08:36:05.515837 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="proxy-httpd" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.515843 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="proxy-httpd" Dec 01 08:36:05 crc kubenswrapper[4744]: E1201 08:36:05.515855 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="sg-core" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.515861 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="sg-core" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.516016 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-notification-agent" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.516027 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="sg-core" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.516037 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="ceilometer-central-agent" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.516047 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a23a18-05c4-409a-950d-0743db8fc652" containerName="proxy-httpd" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.516058 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a458601-7873-4c35-9da5-412411380de8" containerName="dnsmasq-dns" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.517588 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.522866 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.523851 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.557910 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.571597 4744 scope.go:117] "RemoveContainer" containerID="41101381762b71bd360b63ab1e8962d90cd86d9fc2e3a55615047f712ab5c16b" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.637810 4744 scope.go:117] "RemoveContainer" containerID="179e72728e42caf2adc1f8f305005ce810fc06276b7326ef6ea701ce3d37c130" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.666817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.667148 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-scripts\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.667263 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-log-httpd\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.667387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrprg\" (UniqueName: \"kubernetes.io/projected/b8892691-3762-4ddf-a9b4-bcf9417acd02-kube-api-access-zrprg\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.667498 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-config-data\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.667612 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-run-httpd\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.667751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.758001 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-67b74d8fd6-9hczb"] Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.759675 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.763837 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.766670 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769043 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-log-httpd\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrprg\" (UniqueName: \"kubernetes.io/projected/b8892691-3762-4ddf-a9b4-bcf9417acd02-kube-api-access-zrprg\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769127 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-config-data\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-run-httpd\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769201 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-scripts\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.769911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-log-httpd\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.770164 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-run-httpd\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.771642 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67b74d8fd6-9hczb"] Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.774638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.775246 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-config-data\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.776626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.780266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-scripts\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.802085 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrprg\" (UniqueName: \"kubernetes.io/projected/b8892691-3762-4ddf-a9b4-bcf9417acd02-kube-api-access-zrprg\") pod \"ceilometer-0\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.871795 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-config-data\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.872602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-config-data-custom\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.872634 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffp6n\" (UniqueName: \"kubernetes.io/projected/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-kube-api-access-ffp6n\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.872674 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-internal-tls-certs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.872713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-logs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.872790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-public-tls-certs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.872870 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-combined-ca-bundle\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.896512 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.901148 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.976402 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-config-data\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.976527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-config-data-custom\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.976547 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffp6n\" (UniqueName: \"kubernetes.io/projected/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-kube-api-access-ffp6n\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.976574 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-internal-tls-certs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.976604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-logs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.976641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-public-tls-certs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.976682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-combined-ca-bundle\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.978259 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-logs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.985610 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-config-data\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.985785 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-combined-ca-bundle\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:05 crc kubenswrapper[4744]: I1201 08:36:05.998892 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-internal-tls-certs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.002475 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-config-data-custom\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.019594 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-public-tls-certs\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.023941 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffp6n\" (UniqueName: \"kubernetes.io/projected/c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d-kube-api-access-ffp6n\") pod \"barbican-api-67b74d8fd6-9hczb\" (UID: \"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d\") " pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.154831 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.302897 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96a23a18-05c4-409a-950d-0743db8fc652" path="/var/lib/kubelet/pods/96a23a18-05c4-409a-950d-0743db8fc652/volumes" Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.408024 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api-log" containerID="cri-o://5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327" gracePeriod=30 Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.408512 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api" containerID="cri-o://3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e" gracePeriod=30 Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.480982 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:06 crc kubenswrapper[4744]: W1201 08:36:06.486676 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8892691_3762_4ddf_a9b4_bcf9417acd02.slice/crio-1714a0065a9377dcab657d9139f4fadf2fb7a8df587ea498fa109c502c9c7cc7 WatchSource:0}: Error finding container 1714a0065a9377dcab657d9139f4fadf2fb7a8df587ea498fa109c502c9c7cc7: Status 404 returned error can't find the container with id 1714a0065a9377dcab657d9139f4fadf2fb7a8df587ea498fa109c502c9c7cc7 Dec 01 08:36:06 crc kubenswrapper[4744]: I1201 08:36:06.655102 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67b74d8fd6-9hczb"] Dec 01 08:36:06 crc kubenswrapper[4744]: W1201 08:36:06.658843 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc92bfcf4_c9e1_412b_a0e0_2aabeac15d5d.slice/crio-a7f7596ba50f86ef0ca92cb30a309ebbf1b005b266e1c10ed8785cbb37b8df88 WatchSource:0}: Error finding container a7f7596ba50f86ef0ca92cb30a309ebbf1b005b266e1c10ed8785cbb37b8df88: Status 404 returned error can't find the container with id a7f7596ba50f86ef0ca92cb30a309ebbf1b005b266e1c10ed8785cbb37b8df88 Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.045499 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.202222 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnd5c\" (UniqueName: \"kubernetes.io/projected/c61efc8b-a454-412f-bc07-2a734a8bf489-kube-api-access-xnd5c\") pod \"c61efc8b-a454-412f-bc07-2a734a8bf489\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.202307 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data-custom\") pod \"c61efc8b-a454-412f-bc07-2a734a8bf489\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.202491 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c61efc8b-a454-412f-bc07-2a734a8bf489-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c61efc8b-a454-412f-bc07-2a734a8bf489" (UID: "c61efc8b-a454-412f-bc07-2a734a8bf489"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.202446 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c61efc8b-a454-412f-bc07-2a734a8bf489-etc-machine-id\") pod \"c61efc8b-a454-412f-bc07-2a734a8bf489\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.202970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c61efc8b-a454-412f-bc07-2a734a8bf489-logs\") pod \"c61efc8b-a454-412f-bc07-2a734a8bf489\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.203094 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-combined-ca-bundle\") pod \"c61efc8b-a454-412f-bc07-2a734a8bf489\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.203192 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-scripts\") pod \"c61efc8b-a454-412f-bc07-2a734a8bf489\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.203235 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data\") pod \"c61efc8b-a454-412f-bc07-2a734a8bf489\" (UID: \"c61efc8b-a454-412f-bc07-2a734a8bf489\") " Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.203813 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c61efc8b-a454-412f-bc07-2a734a8bf489-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.204594 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c61efc8b-a454-412f-bc07-2a734a8bf489-logs" (OuterVolumeSpecName: "logs") pod "c61efc8b-a454-412f-bc07-2a734a8bf489" (UID: "c61efc8b-a454-412f-bc07-2a734a8bf489"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.212586 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61efc8b-a454-412f-bc07-2a734a8bf489-kube-api-access-xnd5c" (OuterVolumeSpecName: "kube-api-access-xnd5c") pod "c61efc8b-a454-412f-bc07-2a734a8bf489" (UID: "c61efc8b-a454-412f-bc07-2a734a8bf489"). InnerVolumeSpecName "kube-api-access-xnd5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.212819 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c61efc8b-a454-412f-bc07-2a734a8bf489" (UID: "c61efc8b-a454-412f-bc07-2a734a8bf489"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.213005 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-scripts" (OuterVolumeSpecName: "scripts") pod "c61efc8b-a454-412f-bc07-2a734a8bf489" (UID: "c61efc8b-a454-412f-bc07-2a734a8bf489"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.277512 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c61efc8b-a454-412f-bc07-2a734a8bf489" (UID: "c61efc8b-a454-412f-bc07-2a734a8bf489"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.296437 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data" (OuterVolumeSpecName: "config-data") pod "c61efc8b-a454-412f-bc07-2a734a8bf489" (UID: "c61efc8b-a454-412f-bc07-2a734a8bf489"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.306629 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c61efc8b-a454-412f-bc07-2a734a8bf489-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.306664 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.306674 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.306682 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.306690 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnd5c\" (UniqueName: \"kubernetes.io/projected/c61efc8b-a454-412f-bc07-2a734a8bf489-kube-api-access-xnd5c\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.306698 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c61efc8b-a454-412f-bc07-2a734a8bf489-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.417199 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67b74d8fd6-9hczb" event={"ID":"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d","Type":"ContainerStarted","Data":"d40403a19ba65add7365a65c88ab0029afc0a6e494e7ded28031a3006a0666a6"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.417700 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67b74d8fd6-9hczb" event={"ID":"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d","Type":"ContainerStarted","Data":"bcc5318af42d7632ff12e24bcb66b5fe0960bc179179099ac5b4cbcb0810a55e"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.417776 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.417837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67b74d8fd6-9hczb" event={"ID":"c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d","Type":"ContainerStarted","Data":"a7f7596ba50f86ef0ca92cb30a309ebbf1b005b266e1c10ed8785cbb37b8df88"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.417853 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.419369 4744 generic.go:334] "Generic (PLEG): container finished" podID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerID="3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e" exitCode=0 Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.419392 4744 generic.go:334] "Generic (PLEG): container finished" podID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerID="5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327" exitCode=143 Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.419492 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c61efc8b-a454-412f-bc07-2a734a8bf489","Type":"ContainerDied","Data":"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.419529 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c61efc8b-a454-412f-bc07-2a734a8bf489","Type":"ContainerDied","Data":"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.419540 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c61efc8b-a454-412f-bc07-2a734a8bf489","Type":"ContainerDied","Data":"0a261a421293bfdefef719a28b5736ceed1f6ab0216584c73cf416d296df3fbf"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.419555 4744 scope.go:117] "RemoveContainer" containerID="3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.419619 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.421778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerStarted","Data":"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.421818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerStarted","Data":"1714a0065a9377dcab657d9139f4fadf2fb7a8df587ea498fa109c502c9c7cc7"} Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.451955 4744 scope.go:117] "RemoveContainer" containerID="5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.452122 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-67b74d8fd6-9hczb" podStartSLOduration=2.452107126 podStartE2EDuration="2.452107126s" podCreationTimestamp="2025-12-01 08:36:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:07.439727354 +0000 UTC m=+1099.428785285" watchObservedRunningTime="2025-12-01 08:36:07.452107126 +0000 UTC m=+1099.441165047" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.475395 4744 scope.go:117] "RemoveContainer" containerID="3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.476005 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:07 crc kubenswrapper[4744]: E1201 08:36:07.477815 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e\": container with ID starting with 3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e not found: ID does not exist" containerID="3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.477862 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e"} err="failed to get container status \"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e\": rpc error: code = NotFound desc = could not find container \"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e\": container with ID starting with 3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e not found: ID does not exist" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.477895 4744 scope.go:117] "RemoveContainer" containerID="5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327" Dec 01 08:36:07 crc kubenswrapper[4744]: E1201 08:36:07.479302 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327\": container with ID starting with 5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327 not found: ID does not exist" containerID="5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.479369 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327"} err="failed to get container status \"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327\": rpc error: code = NotFound desc = could not find container \"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327\": container with ID starting with 5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327 not found: ID does not exist" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.479402 4744 scope.go:117] "RemoveContainer" containerID="3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.479767 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e"} err="failed to get container status \"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e\": rpc error: code = NotFound desc = could not find container \"3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e\": container with ID starting with 3d32d13b0eabd11bf22cb4f96f805b226668db8ab04bb0e64d335f9b63f2569e not found: ID does not exist" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.479803 4744 scope.go:117] "RemoveContainer" containerID="5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.480069 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327"} err="failed to get container status \"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327\": rpc error: code = NotFound desc = could not find container \"5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327\": container with ID starting with 5b8a6785b2b5596fed3866ac3a76fc046d6739060dfddb526d54d5b54e4a7327 not found: ID does not exist" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.486664 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.500477 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:07 crc kubenswrapper[4744]: E1201 08:36:07.500955 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api-log" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.500973 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api-log" Dec 01 08:36:07 crc kubenswrapper[4744]: E1201 08:36:07.500984 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.500991 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.501220 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api-log" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.501253 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" containerName="cinder-api" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.502342 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.504786 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.504972 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.506211 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.509266 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612050 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612376 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-scripts\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612400 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8rp4\" (UniqueName: \"kubernetes.io/projected/08db63f5-2255-45a5-b2c2-b5baaed9fbca-kube-api-access-m8rp4\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612448 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612490 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/08db63f5-2255-45a5-b2c2-b5baaed9fbca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-public-tls-certs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612664 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-config-data-custom\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612889 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-config-data\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.612933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08db63f5-2255-45a5-b2c2-b5baaed9fbca-logs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714423 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-config-data\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714478 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08db63f5-2255-45a5-b2c2-b5baaed9fbca-logs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714586 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714614 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-scripts\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8rp4\" (UniqueName: \"kubernetes.io/projected/08db63f5-2255-45a5-b2c2-b5baaed9fbca-kube-api-access-m8rp4\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/08db63f5-2255-45a5-b2c2-b5baaed9fbca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-public-tls-certs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714787 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-config-data-custom\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.714982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08db63f5-2255-45a5-b2c2-b5baaed9fbca-logs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.715247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/08db63f5-2255-45a5-b2c2-b5baaed9fbca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.719559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-public-tls-certs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.719711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-scripts\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.719890 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-config-data-custom\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.720721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-config-data\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.722356 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.729126 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08db63f5-2255-45a5-b2c2-b5baaed9fbca-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.732613 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8rp4\" (UniqueName: \"kubernetes.io/projected/08db63f5-2255-45a5-b2c2-b5baaed9fbca-kube-api-access-m8rp4\") pod \"cinder-api-0\" (UID: \"08db63f5-2255-45a5-b2c2-b5baaed9fbca\") " pod="openstack/cinder-api-0" Dec 01 08:36:07 crc kubenswrapper[4744]: I1201 08:36:07.837778 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 08:36:08 crc kubenswrapper[4744]: W1201 08:36:08.285895 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08db63f5_2255_45a5_b2c2_b5baaed9fbca.slice/crio-7f0cf66810629c2c5b53db17c011cb3eb6b51872f88c2e715295252acd12a586 WatchSource:0}: Error finding container 7f0cf66810629c2c5b53db17c011cb3eb6b51872f88c2e715295252acd12a586: Status 404 returned error can't find the container with id 7f0cf66810629c2c5b53db17c011cb3eb6b51872f88c2e715295252acd12a586 Dec 01 08:36:08 crc kubenswrapper[4744]: I1201 08:36:08.318584 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61efc8b-a454-412f-bc07-2a734a8bf489" path="/var/lib/kubelet/pods/c61efc8b-a454-412f-bc07-2a734a8bf489/volumes" Dec 01 08:36:08 crc kubenswrapper[4744]: I1201 08:36:08.319454 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 08:36:08 crc kubenswrapper[4744]: I1201 08:36:08.464002 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerStarted","Data":"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c"} Dec 01 08:36:08 crc kubenswrapper[4744]: I1201 08:36:08.468252 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"08db63f5-2255-45a5-b2c2-b5baaed9fbca","Type":"ContainerStarted","Data":"7f0cf66810629c2c5b53db17c011cb3eb6b51872f88c2e715295252acd12a586"} Dec 01 08:36:09 crc kubenswrapper[4744]: I1201 08:36:09.280915 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:36:09 crc kubenswrapper[4744]: I1201 08:36:09.490247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"08db63f5-2255-45a5-b2c2-b5baaed9fbca","Type":"ContainerStarted","Data":"523e123d0b1f5a9ec39caeb9a880644656e77cb81f761b8113c5497ee152d7a4"} Dec 01 08:36:09 crc kubenswrapper[4744]: I1201 08:36:09.492570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerStarted","Data":"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e"} Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.450585 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.503117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"08db63f5-2255-45a5-b2c2-b5baaed9fbca","Type":"ContainerStarted","Data":"8ac313b275dbad62e21cae37925c4003301007882c531ee7c590b237eb39641f"} Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.504396 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.514645 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.529983 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.529961319 podStartE2EDuration="3.529961319s" podCreationTimestamp="2025-12-01 08:36:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:10.52086639 +0000 UTC m=+1102.509924321" watchObservedRunningTime="2025-12-01 08:36:10.529961319 +0000 UTC m=+1102.519019250" Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.928568 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.975915 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-pjtl9"] Dec 01 08:36:10 crc kubenswrapper[4744]: I1201 08:36:10.976208 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" podUID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerName="dnsmasq-dns" containerID="cri-o://fad23623055fac729576a02472859863ce2d225d63941d7207a0100984ae31de" gracePeriod=10 Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.194590 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.243260 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.515077 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerStarted","Data":"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0"} Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.516241 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.519960 4744 generic.go:334] "Generic (PLEG): container finished" podID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerID="fad23623055fac729576a02472859863ce2d225d63941d7207a0100984ae31de" exitCode=0 Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.520152 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="cinder-scheduler" containerID="cri-o://d55982341b7a8f7f0bde85b82967b57831a2b833e37c1e8b12d618c74b7e4c27" gracePeriod=30 Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.520431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" event={"ID":"3dfafea4-b9db-479a-9876-714ee1dfb14b","Type":"ContainerDied","Data":"fad23623055fac729576a02472859863ce2d225d63941d7207a0100984ae31de"} Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.520461 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" event={"ID":"3dfafea4-b9db-479a-9876-714ee1dfb14b","Type":"ContainerDied","Data":"2d22950627f919a594cbd1634edebb58431eab4776a9834ff3c94e62a9d5c81f"} Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.520472 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d22950627f919a594cbd1634edebb58431eab4776a9834ff3c94e62a9d5c81f" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.520929 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="probe" containerID="cri-o://c8b83fd4f7450a8f5a042a0eeabaf0c5af734c24853a17d7669c43424340ed4c" gracePeriod=30 Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.558068 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.991692016 podStartE2EDuration="6.558043441s" podCreationTimestamp="2025-12-01 08:36:05 +0000 UTC" firstStartedPulling="2025-12-01 08:36:06.488663474 +0000 UTC m=+1098.477721385" lastFinishedPulling="2025-12-01 08:36:11.055014899 +0000 UTC m=+1103.044072810" observedRunningTime="2025-12-01 08:36:11.549093127 +0000 UTC m=+1103.538151058" watchObservedRunningTime="2025-12-01 08:36:11.558043441 +0000 UTC m=+1103.547101362" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.602002 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.709035 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-685cbf69b9-jrc9r" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.709733 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-nb\") pod \"3dfafea4-b9db-479a-9876-714ee1dfb14b\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.709812 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-config\") pod \"3dfafea4-b9db-479a-9876-714ee1dfb14b\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.709869 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-224d8\" (UniqueName: \"kubernetes.io/projected/3dfafea4-b9db-479a-9876-714ee1dfb14b-kube-api-access-224d8\") pod \"3dfafea4-b9db-479a-9876-714ee1dfb14b\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.709932 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-svc\") pod \"3dfafea4-b9db-479a-9876-714ee1dfb14b\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.709971 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-sb\") pod \"3dfafea4-b9db-479a-9876-714ee1dfb14b\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.710065 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-swift-storage-0\") pod \"3dfafea4-b9db-479a-9876-714ee1dfb14b\" (UID: \"3dfafea4-b9db-479a-9876-714ee1dfb14b\") " Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.721681 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dfafea4-b9db-479a-9876-714ee1dfb14b-kube-api-access-224d8" (OuterVolumeSpecName: "kube-api-access-224d8") pod "3dfafea4-b9db-479a-9876-714ee1dfb14b" (UID: "3dfafea4-b9db-479a-9876-714ee1dfb14b"). InnerVolumeSpecName "kube-api-access-224d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.785991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3dfafea4-b9db-479a-9876-714ee1dfb14b" (UID: "3dfafea4-b9db-479a-9876-714ee1dfb14b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.789769 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84cf778f56-stb8l"] Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.791580 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84cf778f56-stb8l" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-api" containerID="cri-o://481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed" gracePeriod=30 Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.792071 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84cf778f56-stb8l" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-httpd" containerID="cri-o://c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547" gracePeriod=30 Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.806238 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3dfafea4-b9db-479a-9876-714ee1dfb14b" (UID: "3dfafea4-b9db-479a-9876-714ee1dfb14b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.824512 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.824854 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.824869 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-224d8\" (UniqueName: \"kubernetes.io/projected/3dfafea4-b9db-479a-9876-714ee1dfb14b-kube-api-access-224d8\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.830941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3dfafea4-b9db-479a-9876-714ee1dfb14b" (UID: "3dfafea4-b9db-479a-9876-714ee1dfb14b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.837002 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3dfafea4-b9db-479a-9876-714ee1dfb14b" (UID: "3dfafea4-b9db-479a-9876-714ee1dfb14b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.862269 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-config" (OuterVolumeSpecName: "config") pod "3dfafea4-b9db-479a-9876-714ee1dfb14b" (UID: "3dfafea4-b9db-479a-9876-714ee1dfb14b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.928588 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.928859 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:11 crc kubenswrapper[4744]: I1201 08:36:11.928925 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dfafea4-b9db-479a-9876-714ee1dfb14b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:12 crc kubenswrapper[4744]: I1201 08:36:12.530366 4744 generic.go:334] "Generic (PLEG): container finished" podID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerID="c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547" exitCode=0 Dec 01 08:36:12 crc kubenswrapper[4744]: I1201 08:36:12.530441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cf778f56-stb8l" event={"ID":"39694e5c-5d4d-4d84-803b-1557e0c49f43","Type":"ContainerDied","Data":"c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547"} Dec 01 08:36:12 crc kubenswrapper[4744]: I1201 08:36:12.532352 4744 generic.go:334] "Generic (PLEG): container finished" podID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerID="c8b83fd4f7450a8f5a042a0eeabaf0c5af734c24853a17d7669c43424340ed4c" exitCode=0 Dec 01 08:36:12 crc kubenswrapper[4744]: I1201 08:36:12.532473 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-pjtl9" Dec 01 08:36:12 crc kubenswrapper[4744]: I1201 08:36:12.532468 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0aa35ddd-0dca-46e2-8f4f-49060a63d45c","Type":"ContainerDied","Data":"c8b83fd4f7450a8f5a042a0eeabaf0c5af734c24853a17d7669c43424340ed4c"} Dec 01 08:36:12 crc kubenswrapper[4744]: I1201 08:36:12.559357 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-pjtl9"] Dec 01 08:36:12 crc kubenswrapper[4744]: I1201 08:36:12.569144 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-pjtl9"] Dec 01 08:36:13 crc kubenswrapper[4744]: I1201 08:36:13.124744 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:14 crc kubenswrapper[4744]: I1201 08:36:14.296555 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dfafea4-b9db-479a-9876-714ee1dfb14b" path="/var/lib/kubelet/pods/3dfafea4-b9db-479a-9876-714ee1dfb14b/volumes" Dec 01 08:36:14 crc kubenswrapper[4744]: I1201 08:36:14.519800 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67b74d8fd6-9hczb" Dec 01 08:36:14 crc kubenswrapper[4744]: I1201 08:36:14.596704 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-556674cd78-bw7k7"] Dec 01 08:36:14 crc kubenswrapper[4744]: I1201 08:36:14.597043 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-556674cd78-bw7k7" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api-log" containerID="cri-o://59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65" gracePeriod=30 Dec 01 08:36:14 crc kubenswrapper[4744]: I1201 08:36:14.597311 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-556674cd78-bw7k7" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api" containerID="cri-o://5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3" gracePeriod=30 Dec 01 08:36:14 crc kubenswrapper[4744]: I1201 08:36:14.615634 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556674cd78-bw7k7" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": EOF" Dec 01 08:36:15 crc kubenswrapper[4744]: I1201 08:36:15.600909 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerID="59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65" exitCode=143 Dec 01 08:36:15 crc kubenswrapper[4744]: I1201 08:36:15.600984 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556674cd78-bw7k7" event={"ID":"f1473d13-7648-4ae3-8919-0bd62c6893e8","Type":"ContainerDied","Data":"59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65"} Dec 01 08:36:15 crc kubenswrapper[4744]: I1201 08:36:15.610351 4744 generic.go:334] "Generic (PLEG): container finished" podID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerID="d55982341b7a8f7f0bde85b82967b57831a2b833e37c1e8b12d618c74b7e4c27" exitCode=0 Dec 01 08:36:15 crc kubenswrapper[4744]: I1201 08:36:15.610426 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0aa35ddd-0dca-46e2-8f4f-49060a63d45c","Type":"ContainerDied","Data":"d55982341b7a8f7f0bde85b82967b57831a2b833e37c1e8b12d618c74b7e4c27"} Dec 01 08:36:15 crc kubenswrapper[4744]: I1201 08:36:15.783519 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-77497cdd8f-s4w2q" Dec 01 08:36:15 crc kubenswrapper[4744]: I1201 08:36:15.913776 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86964bb7b8-4bgsm" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.004959 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.108889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data-custom\") pod \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.108962 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr22t\" (UniqueName: \"kubernetes.io/projected/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-kube-api-access-pr22t\") pod \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.109044 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-etc-machine-id\") pod \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.109213 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-scripts\") pod \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.109244 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data\") pod \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.109276 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-combined-ca-bundle\") pod \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\" (UID: \"0aa35ddd-0dca-46e2-8f4f-49060a63d45c\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.109513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0aa35ddd-0dca-46e2-8f4f-49060a63d45c" (UID: "0aa35ddd-0dca-46e2-8f4f-49060a63d45c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.110311 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.128232 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-kube-api-access-pr22t" (OuterVolumeSpecName: "kube-api-access-pr22t") pod "0aa35ddd-0dca-46e2-8f4f-49060a63d45c" (UID: "0aa35ddd-0dca-46e2-8f4f-49060a63d45c"). InnerVolumeSpecName "kube-api-access-pr22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.138602 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0aa35ddd-0dca-46e2-8f4f-49060a63d45c" (UID: "0aa35ddd-0dca-46e2-8f4f-49060a63d45c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.142872 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-scripts" (OuterVolumeSpecName: "scripts") pod "0aa35ddd-0dca-46e2-8f4f-49060a63d45c" (UID: "0aa35ddd-0dca-46e2-8f4f-49060a63d45c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.211870 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr22t\" (UniqueName: \"kubernetes.io/projected/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-kube-api-access-pr22t\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.211901 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.211912 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.213326 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0aa35ddd-0dca-46e2-8f4f-49060a63d45c" (UID: "0aa35ddd-0dca-46e2-8f4f-49060a63d45c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.270230 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data" (OuterVolumeSpecName: "config-data") pod "0aa35ddd-0dca-46e2-8f4f-49060a63d45c" (UID: "0aa35ddd-0dca-46e2-8f4f-49060a63d45c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.311496 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.314760 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.314951 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aa35ddd-0dca-46e2-8f4f-49060a63d45c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.415815 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-config\") pod \"39694e5c-5d4d-4d84-803b-1557e0c49f43\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.415883 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-combined-ca-bundle\") pod \"39694e5c-5d4d-4d84-803b-1557e0c49f43\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.415915 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-httpd-config\") pod \"39694e5c-5d4d-4d84-803b-1557e0c49f43\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.415949 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-ovndb-tls-certs\") pod \"39694e5c-5d4d-4d84-803b-1557e0c49f43\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.415968 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x6jn\" (UniqueName: \"kubernetes.io/projected/39694e5c-5d4d-4d84-803b-1557e0c49f43-kube-api-access-9x6jn\") pod \"39694e5c-5d4d-4d84-803b-1557e0c49f43\" (UID: \"39694e5c-5d4d-4d84-803b-1557e0c49f43\") " Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.420601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "39694e5c-5d4d-4d84-803b-1557e0c49f43" (UID: "39694e5c-5d4d-4d84-803b-1557e0c49f43"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.427613 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39694e5c-5d4d-4d84-803b-1557e0c49f43-kube-api-access-9x6jn" (OuterVolumeSpecName: "kube-api-access-9x6jn") pod "39694e5c-5d4d-4d84-803b-1557e0c49f43" (UID: "39694e5c-5d4d-4d84-803b-1557e0c49f43"). InnerVolumeSpecName "kube-api-access-9x6jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.487562 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39694e5c-5d4d-4d84-803b-1557e0c49f43" (UID: "39694e5c-5d4d-4d84-803b-1557e0c49f43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.493154 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-config" (OuterVolumeSpecName: "config") pod "39694e5c-5d4d-4d84-803b-1557e0c49f43" (UID: "39694e5c-5d4d-4d84-803b-1557e0c49f43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.496708 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "39694e5c-5d4d-4d84-803b-1557e0c49f43" (UID: "39694e5c-5d4d-4d84-803b-1557e0c49f43"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.520696 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.521058 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.521156 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.521255 4744 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39694e5c-5d4d-4d84-803b-1557e0c49f43-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.521389 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x6jn\" (UniqueName: \"kubernetes.io/projected/39694e5c-5d4d-4d84-803b-1557e0c49f43-kube-api-access-9x6jn\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.622311 4744 generic.go:334] "Generic (PLEG): container finished" podID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerID="481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed" exitCode=0 Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.622363 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cf778f56-stb8l" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.622396 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cf778f56-stb8l" event={"ID":"39694e5c-5d4d-4d84-803b-1557e0c49f43","Type":"ContainerDied","Data":"481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed"} Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.622520 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cf778f56-stb8l" event={"ID":"39694e5c-5d4d-4d84-803b-1557e0c49f43","Type":"ContainerDied","Data":"822c2216940c509513776fab6e7a8be9d828a5084cdad1945db6d0cc7e1c86d4"} Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.622556 4744 scope.go:117] "RemoveContainer" containerID="c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.627450 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0aa35ddd-0dca-46e2-8f4f-49060a63d45c","Type":"ContainerDied","Data":"7be580839ea42e8ed7878d1e6b6936ce0ddbc1f91772ad1cd6f06d7ac17bd19d"} Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.627554 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.649121 4744 scope.go:117] "RemoveContainer" containerID="481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.663262 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84cf778f56-stb8l"] Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.675186 4744 scope.go:117] "RemoveContainer" containerID="c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547" Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.675619 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547\": container with ID starting with c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547 not found: ID does not exist" containerID="c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.675649 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547"} err="failed to get container status \"c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547\": rpc error: code = NotFound desc = could not find container \"c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547\": container with ID starting with c837ec622c23d437db774d2d0c169ac110dc591d1df0cf7f6edb98fc6fef5547 not found: ID does not exist" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.675670 4744 scope.go:117] "RemoveContainer" containerID="481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.675714 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-84cf778f56-stb8l"] Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.676102 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed\": container with ID starting with 481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed not found: ID does not exist" containerID="481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.676158 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed"} err="failed to get container status \"481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed\": rpc error: code = NotFound desc = could not find container \"481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed\": container with ID starting with 481a6a9edd225e5c9f7a6ba936a6e082a5a3a9720d172c2db9d45ac352ca1fed not found: ID does not exist" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.676199 4744 scope.go:117] "RemoveContainer" containerID="c8b83fd4f7450a8f5a042a0eeabaf0c5af734c24853a17d7669c43424340ed4c" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.686536 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.700111 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.719277 4744 scope.go:117] "RemoveContainer" containerID="d55982341b7a8f7f0bde85b82967b57831a2b833e37c1e8b12d618c74b7e4c27" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.722371 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.722749 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-httpd" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.722764 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-httpd" Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.722788 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="probe" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.722797 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="probe" Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.722820 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="cinder-scheduler" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.722829 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="cinder-scheduler" Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.722840 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerName="init" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.722848 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerName="init" Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.722861 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerName="dnsmasq-dns" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.722866 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerName="dnsmasq-dns" Dec 01 08:36:16 crc kubenswrapper[4744]: E1201 08:36:16.722884 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-api" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.722891 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-api" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.723047 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="cinder-scheduler" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.723062 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-api" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.723072 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dfafea4-b9db-479a-9876-714ee1dfb14b" containerName="dnsmasq-dns" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.723087 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" containerName="neutron-httpd" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.723102 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" containerName="probe" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.724179 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.726432 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.742219 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.929578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-scripts\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.929647 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-config-data\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.929710 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.929746 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j5bc\" (UniqueName: \"kubernetes.io/projected/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-kube-api-access-8j5bc\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.929937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:16 crc kubenswrapper[4744]: I1201 08:36:16.929975 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.031937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j5bc\" (UniqueName: \"kubernetes.io/projected/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-kube-api-access-8j5bc\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.032021 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.032057 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.032117 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-scripts\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.032174 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-config-data\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.032216 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.032317 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.037203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.038985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-scripts\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.041448 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-config-data\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.058265 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.062874 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j5bc\" (UniqueName: \"kubernetes.io/projected/23516ec1-d13c-4dfd-b20f-207ba2bb0d26-kube-api-access-8j5bc\") pod \"cinder-scheduler-0\" (UID: \"23516ec1-d13c-4dfd-b20f-207ba2bb0d26\") " pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.343992 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 08:36:17 crc kubenswrapper[4744]: I1201 08:36:17.814657 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 08:36:17 crc kubenswrapper[4744]: W1201 08:36:17.819026 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23516ec1_d13c_4dfd_b20f_207ba2bb0d26.slice/crio-4ecb55ad1f8b6b13c8e2cc7375654a91e391f68f4d25cf93ccd499fb51159d1a WatchSource:0}: Error finding container 4ecb55ad1f8b6b13c8e2cc7375654a91e391f68f4d25cf93ccd499fb51159d1a: Status 404 returned error can't find the container with id 4ecb55ad1f8b6b13c8e2cc7375654a91e391f68f4d25cf93ccd499fb51159d1a Dec 01 08:36:18 crc kubenswrapper[4744]: I1201 08:36:18.294298 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa35ddd-0dca-46e2-8f4f-49060a63d45c" path="/var/lib/kubelet/pods/0aa35ddd-0dca-46e2-8f4f-49060a63d45c/volumes" Dec 01 08:36:18 crc kubenswrapper[4744]: I1201 08:36:18.295152 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39694e5c-5d4d-4d84-803b-1557e0c49f43" path="/var/lib/kubelet/pods/39694e5c-5d4d-4d84-803b-1557e0c49f43/volumes" Dec 01 08:36:18 crc kubenswrapper[4744]: I1201 08:36:18.662035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23516ec1-d13c-4dfd-b20f-207ba2bb0d26","Type":"ContainerStarted","Data":"c5d30597d3333154d8fec90427277932342e732e5b215ebfd33fea59c1072d64"} Dec 01 08:36:18 crc kubenswrapper[4744]: I1201 08:36:18.662376 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23516ec1-d13c-4dfd-b20f-207ba2bb0d26","Type":"ContainerStarted","Data":"4ecb55ad1f8b6b13c8e2cc7375654a91e391f68f4d25cf93ccd499fb51159d1a"} Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.083311 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556674cd78-bw7k7" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:51984->10.217.0.159:9311: read: connection reset by peer" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.083822 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556674cd78-bw7k7" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:51996->10.217.0.159:9311: read: connection reset by peer" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.512567 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.672024 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerID="5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3" exitCode=0 Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.672099 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556674cd78-bw7k7" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.672106 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556674cd78-bw7k7" event={"ID":"f1473d13-7648-4ae3-8919-0bd62c6893e8","Type":"ContainerDied","Data":"5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3"} Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.672639 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556674cd78-bw7k7" event={"ID":"f1473d13-7648-4ae3-8919-0bd62c6893e8","Type":"ContainerDied","Data":"ca16c627673803853a3a62edc46b7813faa120011f502dc6d4e56093a4358a78"} Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.672665 4744 scope.go:117] "RemoveContainer" containerID="5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.678116 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23516ec1-d13c-4dfd-b20f-207ba2bb0d26","Type":"ContainerStarted","Data":"213605d8b03521b17107a29bbf2acffc60e7b9ae61b5ec1c929484833938d94a"} Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.684250 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-combined-ca-bundle\") pod \"f1473d13-7648-4ae3-8919-0bd62c6893e8\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.684719 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1473d13-7648-4ae3-8919-0bd62c6893e8-logs\") pod \"f1473d13-7648-4ae3-8919-0bd62c6893e8\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.684953 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data-custom\") pod \"f1473d13-7648-4ae3-8919-0bd62c6893e8\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.684979 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data\") pod \"f1473d13-7648-4ae3-8919-0bd62c6893e8\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.685104 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1473d13-7648-4ae3-8919-0bd62c6893e8-logs" (OuterVolumeSpecName: "logs") pod "f1473d13-7648-4ae3-8919-0bd62c6893e8" (UID: "f1473d13-7648-4ae3-8919-0bd62c6893e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.685218 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrbcz\" (UniqueName: \"kubernetes.io/projected/f1473d13-7648-4ae3-8919-0bd62c6893e8-kube-api-access-mrbcz\") pod \"f1473d13-7648-4ae3-8919-0bd62c6893e8\" (UID: \"f1473d13-7648-4ae3-8919-0bd62c6893e8\") " Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.688006 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1473d13-7648-4ae3-8919-0bd62c6893e8-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.692669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f1473d13-7648-4ae3-8919-0bd62c6893e8" (UID: "f1473d13-7648-4ae3-8919-0bd62c6893e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.694955 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1473d13-7648-4ae3-8919-0bd62c6893e8-kube-api-access-mrbcz" (OuterVolumeSpecName: "kube-api-access-mrbcz") pod "f1473d13-7648-4ae3-8919-0bd62c6893e8" (UID: "f1473d13-7648-4ae3-8919-0bd62c6893e8"). InnerVolumeSpecName "kube-api-access-mrbcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.720675 4744 scope.go:117] "RemoveContainer" containerID="59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.723123 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1473d13-7648-4ae3-8919-0bd62c6893e8" (UID: "f1473d13-7648-4ae3-8919-0bd62c6893e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.753171 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data" (OuterVolumeSpecName: "config-data") pod "f1473d13-7648-4ae3-8919-0bd62c6893e8" (UID: "f1473d13-7648-4ae3-8919-0bd62c6893e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.791729 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.791771 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.791784 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrbcz\" (UniqueName: \"kubernetes.io/projected/f1473d13-7648-4ae3-8919-0bd62c6893e8-kube-api-access-mrbcz\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.791796 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1473d13-7648-4ae3-8919-0bd62c6893e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.797118 4744 scope.go:117] "RemoveContainer" containerID="5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3" Dec 01 08:36:19 crc kubenswrapper[4744]: E1201 08:36:19.797621 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3\": container with ID starting with 5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3 not found: ID does not exist" containerID="5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.797769 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3"} err="failed to get container status \"5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3\": rpc error: code = NotFound desc = could not find container \"5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3\": container with ID starting with 5179fd03656b03243d9146cb5881f620959b9ab1b6c93aafab77a8100d7073c3 not found: ID does not exist" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.797895 4744 scope.go:117] "RemoveContainer" containerID="59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65" Dec 01 08:36:19 crc kubenswrapper[4744]: E1201 08:36:19.800287 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65\": container with ID starting with 59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65 not found: ID does not exist" containerID="59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.800335 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65"} err="failed to get container status \"59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65\": rpc error: code = NotFound desc = could not find container \"59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65\": container with ID starting with 59b8225a0318bf086092fc61abb0cdfcc5ecd62e284fe040f9d9bdd84b683f65 not found: ID does not exist" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.805540 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.805523876 podStartE2EDuration="3.805523876s" podCreationTimestamp="2025-12-01 08:36:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:19.721202325 +0000 UTC m=+1111.710260256" watchObservedRunningTime="2025-12-01 08:36:19.805523876 +0000 UTC m=+1111.794581797" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.805794 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 08:36:19 crc kubenswrapper[4744]: E1201 08:36:19.806119 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.806135 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api" Dec 01 08:36:19 crc kubenswrapper[4744]: E1201 08:36:19.806155 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api-log" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.806162 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api-log" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.806321 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.806342 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" containerName="barbican-api-log" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.806868 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.808365 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.809554 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-wj5q9" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.809715 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.826110 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.893133 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.893199 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmnjj\" (UniqueName: \"kubernetes.io/projected/181b6931-2498-4979-a9b1-f6cbc925c78e-kube-api-access-rmnjj\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.893282 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.893304 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config-secret\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.971188 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.994820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.994918 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmnjj\" (UniqueName: \"kubernetes.io/projected/181b6931-2498-4979-a9b1-f6cbc925c78e-kube-api-access-rmnjj\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.995032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.995059 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config-secret\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.995962 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:19 crc kubenswrapper[4744]: I1201 08:36:19.998335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config-secret\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.001997 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.025961 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmnjj\" (UniqueName: \"kubernetes.io/projected/181b6931-2498-4979-a9b1-f6cbc925c78e-kube-api-access-rmnjj\") pod \"openstackclient\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " pod="openstack/openstackclient" Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.046899 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-556674cd78-bw7k7"] Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.054987 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-556674cd78-bw7k7"] Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.126385 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.304545 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1473d13-7648-4ae3-8919-0bd62c6893e8" path="/var/lib/kubelet/pods/f1473d13-7648-4ae3-8919-0bd62c6893e8/volumes" Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.665670 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 08:36:20 crc kubenswrapper[4744]: W1201 08:36:20.666359 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod181b6931_2498_4979_a9b1_f6cbc925c78e.slice/crio-007e47a8ce4ec48b07f31cd2b7ffd782676bc8c5eb9d8fdb0eb3fdcde8484242 WatchSource:0}: Error finding container 007e47a8ce4ec48b07f31cd2b7ffd782676bc8c5eb9d8fdb0eb3fdcde8484242: Status 404 returned error can't find the container with id 007e47a8ce4ec48b07f31cd2b7ffd782676bc8c5eb9d8fdb0eb3fdcde8484242 Dec 01 08:36:20 crc kubenswrapper[4744]: I1201 08:36:20.689617 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"181b6931-2498-4979-a9b1-f6cbc925c78e","Type":"ContainerStarted","Data":"007e47a8ce4ec48b07f31cd2b7ffd782676bc8c5eb9d8fdb0eb3fdcde8484242"} Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.346058 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.707929 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-696c89c747-77nz8"] Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.709428 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.711695 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.711904 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.714000 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.733669 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-696c89c747-77nz8"] Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748280 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-combined-ca-bundle\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748356 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-internal-tls-certs\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-run-httpd\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-public-tls-certs\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-log-httpd\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-config-data\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7zsq\" (UniqueName: \"kubernetes.io/projected/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-kube-api-access-l7zsq\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.748769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-etc-swift\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849452 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-internal-tls-certs\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-run-httpd\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849582 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-public-tls-certs\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-log-httpd\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-config-data\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849660 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7zsq\" (UniqueName: \"kubernetes.io/projected/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-kube-api-access-l7zsq\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849679 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-etc-swift\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.849723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-combined-ca-bundle\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.850271 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-run-httpd\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.850686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-log-httpd\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.857541 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-combined-ca-bundle\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.857734 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-internal-tls-certs\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.859327 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-etc-swift\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.865149 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-config-data\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.865672 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-public-tls-certs\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:22 crc kubenswrapper[4744]: I1201 08:36:22.870173 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7zsq\" (UniqueName: \"kubernetes.io/projected/2e5f2594-2a7b-4c08-ac37-3e258f99fbb4-kube-api-access-l7zsq\") pod \"swift-proxy-696c89c747-77nz8\" (UID: \"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4\") " pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.035026 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.603111 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-696c89c747-77nz8"] Dec 01 08:36:23 crc kubenswrapper[4744]: W1201 08:36:23.614987 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e5f2594_2a7b_4c08_ac37_3e258f99fbb4.slice/crio-15502997e4ceb06b874dcc521d6263967a8a11d966d39ceb892fb13c7a8a0764 WatchSource:0}: Error finding container 15502997e4ceb06b874dcc521d6263967a8a11d966d39ceb892fb13c7a8a0764: Status 404 returned error can't find the container with id 15502997e4ceb06b874dcc521d6263967a8a11d966d39ceb892fb13c7a8a0764 Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.710286 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5b7d595658-kpqqh"] Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.712454 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.718472 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.718695 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.718842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-qjzlh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.725131 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5b7d595658-kpqqh"] Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.757522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696c89c747-77nz8" event={"ID":"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4","Type":"ContainerStarted","Data":"15502997e4ceb06b874dcc521d6263967a8a11d966d39ceb892fb13c7a8a0764"} Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.763172 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-combined-ca-bundle\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.763252 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data-custom\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.763288 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.763382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2btk\" (UniqueName: \"kubernetes.io/projected/947af6f3-b9a2-46da-98d3-bed308f4ba29-kube-api-access-v2btk\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.839573 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-zqdmr"] Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.841638 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.865089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.865382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.865546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-svc\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.865644 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.865743 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2btk\" (UniqueName: \"kubernetes.io/projected/947af6f3-b9a2-46da-98d3-bed308f4ba29-kube-api-access-v2btk\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.865848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp5pt\" (UniqueName: \"kubernetes.io/projected/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-kube-api-access-sp5pt\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.866001 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.866161 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-combined-ca-bundle\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.866300 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data-custom\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.866394 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-config\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.874565 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data-custom\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.874798 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-zqdmr"] Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.877483 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.897167 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-combined-ca-bundle\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.898886 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5d5b458c99-8sfkr"] Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.900466 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.901132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2btk\" (UniqueName: \"kubernetes.io/projected/947af6f3-b9a2-46da-98d3-bed308f4ba29-kube-api-access-v2btk\") pod \"heat-engine-5b7d595658-kpqqh\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.923867 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.928734 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5d5b458c99-8sfkr"] Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.983932 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.984034 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-config\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.984092 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.984113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-svc\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.984131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.984166 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp5pt\" (UniqueName: \"kubernetes.io/projected/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-kube-api-access-sp5pt\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.986168 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.986749 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-config\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.987289 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.987829 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-svc\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:23 crc kubenswrapper[4744]: I1201 08:36:23.988358 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.006676 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp5pt\" (UniqueName: \"kubernetes.io/projected/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-kube-api-access-sp5pt\") pod \"dnsmasq-dns-7d978555f9-zqdmr\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.010951 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-869bdb8dbc-7n726"] Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.012694 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.015534 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.034475 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-869bdb8dbc-7n726"] Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.049692 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.089226 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-combined-ca-bundle\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.089568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data-custom\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.089701 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7ljh\" (UniqueName: \"kubernetes.io/projected/33b049d4-4934-46bd-8dff-cf9995702c8d-kube-api-access-f7ljh\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.089871 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-combined-ca-bundle\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.090022 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.090159 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data-custom\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.090269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.090446 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76gd7\" (UniqueName: \"kubernetes.io/projected/3b626810-3162-42db-83b2-9594a48dd935-kube-api-access-76gd7\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.192962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76gd7\" (UniqueName: \"kubernetes.io/projected/3b626810-3162-42db-83b2-9594a48dd935-kube-api-access-76gd7\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.193240 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-combined-ca-bundle\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.193285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data-custom\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.193305 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7ljh\" (UniqueName: \"kubernetes.io/projected/33b049d4-4934-46bd-8dff-cf9995702c8d-kube-api-access-f7ljh\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.193324 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-combined-ca-bundle\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.193351 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.193387 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data-custom\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.193424 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.196910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-combined-ca-bundle\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.199608 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.203075 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data-custom\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.206567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.207156 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-combined-ca-bundle\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.210427 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data-custom\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.217870 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76gd7\" (UniqueName: \"kubernetes.io/projected/3b626810-3162-42db-83b2-9594a48dd935-kube-api-access-76gd7\") pod \"heat-api-5d5b458c99-8sfkr\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.235018 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7ljh\" (UniqueName: \"kubernetes.io/projected/33b049d4-4934-46bd-8dff-cf9995702c8d-kube-api-access-f7ljh\") pod \"heat-cfnapi-869bdb8dbc-7n726\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.301955 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.357117 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.368523 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.429105 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.429650 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-central-agent" containerID="cri-o://e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9" gracePeriod=30 Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.431201 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="proxy-httpd" containerID="cri-o://9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0" gracePeriod=30 Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.431288 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="sg-core" containerID="cri-o://ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e" gracePeriod=30 Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.431332 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-notification-agent" containerID="cri-o://9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c" gracePeriod=30 Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.442493 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.713546 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5b7d595658-kpqqh"] Dec 01 08:36:24 crc kubenswrapper[4744]: W1201 08:36:24.740482 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod947af6f3_b9a2_46da_98d3_bed308f4ba29.slice/crio-ee6e4ca6501a0f464e2478fcf8f0d647b92ff6c43b610d76cfa1011c67fe9758 WatchSource:0}: Error finding container ee6e4ca6501a0f464e2478fcf8f0d647b92ff6c43b610d76cfa1011c67fe9758: Status 404 returned error can't find the container with id ee6e4ca6501a0f464e2478fcf8f0d647b92ff6c43b610d76cfa1011c67fe9758 Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.774694 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5b7d595658-kpqqh" event={"ID":"947af6f3-b9a2-46da-98d3-bed308f4ba29","Type":"ContainerStarted","Data":"ee6e4ca6501a0f464e2478fcf8f0d647b92ff6c43b610d76cfa1011c67fe9758"} Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.783007 4744 generic.go:334] "Generic (PLEG): container finished" podID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerID="9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0" exitCode=0 Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.783041 4744 generic.go:334] "Generic (PLEG): container finished" podID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerID="ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e" exitCode=2 Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.783085 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerDied","Data":"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0"} Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.783111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerDied","Data":"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e"} Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.786097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696c89c747-77nz8" event={"ID":"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4","Type":"ContainerStarted","Data":"24d5c2207cdd4eb303e8df25a8dc063faa16754157b49d3cf63a4807df5714c0"} Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.786135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696c89c747-77nz8" event={"ID":"2e5f2594-2a7b-4c08-ac37-3e258f99fbb4","Type":"ContainerStarted","Data":"7c6ffdce83877584778245ae60156bc2a7e5564c8cf0632b8fa264a50e8dfdb3"} Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.787288 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.787323 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.935551 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-696c89c747-77nz8" podStartSLOduration=2.935531698 podStartE2EDuration="2.935531698s" podCreationTimestamp="2025-12-01 08:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:24.831797845 +0000 UTC m=+1116.820855766" watchObservedRunningTime="2025-12-01 08:36:24.935531698 +0000 UTC m=+1116.924589639" Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.938911 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-zqdmr"] Dec 01 08:36:24 crc kubenswrapper[4744]: I1201 08:36:24.999791 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5d5b458c99-8sfkr"] Dec 01 08:36:25 crc kubenswrapper[4744]: W1201 08:36:25.039790 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b626810_3162_42db_83b2_9594a48dd935.slice/crio-57e9d21ed6a1b258b7c46f35706d343bb4a78c23b2ea8b06d09edcaa3b54736a WatchSource:0}: Error finding container 57e9d21ed6a1b258b7c46f35706d343bb4a78c23b2ea8b06d09edcaa3b54736a: Status 404 returned error can't find the container with id 57e9d21ed6a1b258b7c46f35706d343bb4a78c23b2ea8b06d09edcaa3b54736a Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.144336 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-869bdb8dbc-7n726"] Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.450015 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629519 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-run-httpd\") pod \"b8892691-3762-4ddf-a9b4-bcf9417acd02\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629600 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-scripts\") pod \"b8892691-3762-4ddf-a9b4-bcf9417acd02\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629668 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-log-httpd\") pod \"b8892691-3762-4ddf-a9b4-bcf9417acd02\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629720 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrprg\" (UniqueName: \"kubernetes.io/projected/b8892691-3762-4ddf-a9b4-bcf9417acd02-kube-api-access-zrprg\") pod \"b8892691-3762-4ddf-a9b4-bcf9417acd02\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-sg-core-conf-yaml\") pod \"b8892691-3762-4ddf-a9b4-bcf9417acd02\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-config-data\") pod \"b8892691-3762-4ddf-a9b4-bcf9417acd02\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629885 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-combined-ca-bundle\") pod \"b8892691-3762-4ddf-a9b4-bcf9417acd02\" (UID: \"b8892691-3762-4ddf-a9b4-bcf9417acd02\") " Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.629958 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b8892691-3762-4ddf-a9b4-bcf9417acd02" (UID: "b8892691-3762-4ddf-a9b4-bcf9417acd02"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.630216 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.631000 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b8892691-3762-4ddf-a9b4-bcf9417acd02" (UID: "b8892691-3762-4ddf-a9b4-bcf9417acd02"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.634552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-scripts" (OuterVolumeSpecName: "scripts") pod "b8892691-3762-4ddf-a9b4-bcf9417acd02" (UID: "b8892691-3762-4ddf-a9b4-bcf9417acd02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.645295 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8892691-3762-4ddf-a9b4-bcf9417acd02-kube-api-access-zrprg" (OuterVolumeSpecName: "kube-api-access-zrprg") pod "b8892691-3762-4ddf-a9b4-bcf9417acd02" (UID: "b8892691-3762-4ddf-a9b4-bcf9417acd02"). InnerVolumeSpecName "kube-api-access-zrprg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.665910 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b8892691-3762-4ddf-a9b4-bcf9417acd02" (UID: "b8892691-3762-4ddf-a9b4-bcf9417acd02"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.732881 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.733952 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8892691-3762-4ddf-a9b4-bcf9417acd02-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.734022 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrprg\" (UniqueName: \"kubernetes.io/projected/b8892691-3762-4ddf-a9b4-bcf9417acd02-kube-api-access-zrprg\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.734079 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.752847 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8892691-3762-4ddf-a9b4-bcf9417acd02" (UID: "b8892691-3762-4ddf-a9b4-bcf9417acd02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.814578 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" event={"ID":"33b049d4-4934-46bd-8dff-cf9995702c8d","Type":"ContainerStarted","Data":"4bee75623a79eae1198797b9bc35e43f81d3a71333d02af1e06c13f3d0ae0252"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.833998 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5b7d595658-kpqqh" event={"ID":"947af6f3-b9a2-46da-98d3-bed308f4ba29","Type":"ContainerStarted","Data":"a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.834056 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.835523 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.857453 4744 generic.go:334] "Generic (PLEG): container finished" podID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerID="9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c" exitCode=0 Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.857478 4744 generic.go:334] "Generic (PLEG): container finished" podID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerID="e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9" exitCode=0 Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.857516 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerDied","Data":"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.857541 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerDied","Data":"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.857553 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8892691-3762-4ddf-a9b4-bcf9417acd02","Type":"ContainerDied","Data":"1714a0065a9377dcab657d9139f4fadf2fb7a8df587ea498fa109c502c9c7cc7"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.857568 4744 scope.go:117] "RemoveContainer" containerID="9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.857701 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.861583 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d5b458c99-8sfkr" event={"ID":"3b626810-3162-42db-83b2-9594a48dd935","Type":"ContainerStarted","Data":"57e9d21ed6a1b258b7c46f35706d343bb4a78c23b2ea8b06d09edcaa3b54736a"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.871464 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-config-data" (OuterVolumeSpecName: "config-data") pod "b8892691-3762-4ddf-a9b4-bcf9417acd02" (UID: "b8892691-3762-4ddf-a9b4-bcf9417acd02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.884328 4744 generic.go:334] "Generic (PLEG): container finished" podID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerID="cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac" exitCode=0 Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.885597 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" event={"ID":"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb","Type":"ContainerDied","Data":"cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.885634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" event={"ID":"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb","Type":"ContainerStarted","Data":"4996bc9300e8487e347311584a654f937b8cbada6f45e708d70b201ba7aede5a"} Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.949825 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5b7d595658-kpqqh" podStartSLOduration=2.9497898769999997 podStartE2EDuration="2.949789877s" podCreationTimestamp="2025-12-01 08:36:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:25.851285072 +0000 UTC m=+1117.840342993" watchObservedRunningTime="2025-12-01 08:36:25.949789877 +0000 UTC m=+1117.938847798" Dec 01 08:36:25 crc kubenswrapper[4744]: I1201 08:36:25.955439 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8892691-3762-4ddf-a9b4-bcf9417acd02-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.120244 4744 scope.go:117] "RemoveContainer" containerID="ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.144826 4744 scope.go:117] "RemoveContainer" containerID="9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.190113 4744 scope.go:117] "RemoveContainer" containerID="e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.206751 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.219562 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230073 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.230463 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="proxy-httpd" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230478 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="proxy-httpd" Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.230500 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-central-agent" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230507 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-central-agent" Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.230525 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-notification-agent" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230530 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-notification-agent" Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.230541 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="sg-core" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230546 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="sg-core" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230703 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="proxy-httpd" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230716 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="sg-core" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230731 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-notification-agent" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.230744 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" containerName="ceilometer-central-agent" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.237755 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.237936 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.244634 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.244853 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.248558 4744 scope.go:117] "RemoveContainer" containerID="9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0" Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.249888 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0\": container with ID starting with 9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0 not found: ID does not exist" containerID="9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.250014 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0"} err="failed to get container status \"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0\": rpc error: code = NotFound desc = could not find container \"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0\": container with ID starting with 9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0 not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.250117 4744 scope.go:117] "RemoveContainer" containerID="ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e" Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.250743 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e\": container with ID starting with ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e not found: ID does not exist" containerID="ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.250835 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e"} err="failed to get container status \"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e\": rpc error: code = NotFound desc = could not find container \"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e\": container with ID starting with ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.251232 4744 scope.go:117] "RemoveContainer" containerID="9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c" Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.252820 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c\": container with ID starting with 9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c not found: ID does not exist" containerID="9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.252859 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c"} err="failed to get container status \"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c\": rpc error: code = NotFound desc = could not find container \"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c\": container with ID starting with 9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.252887 4744 scope.go:117] "RemoveContainer" containerID="e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9" Dec 01 08:36:26 crc kubenswrapper[4744]: E1201 08:36:26.253837 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9\": container with ID starting with e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9 not found: ID does not exist" containerID="e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.253862 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9"} err="failed to get container status \"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9\": rpc error: code = NotFound desc = could not find container \"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9\": container with ID starting with e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9 not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.253875 4744 scope.go:117] "RemoveContainer" containerID="9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.254263 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0"} err="failed to get container status \"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0\": rpc error: code = NotFound desc = could not find container \"9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0\": container with ID starting with 9bdb4a6db0e2258ca5288eda5e06ebcdae8f6fce3fd5c6944cea0567dd20d9f0 not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.254283 4744 scope.go:117] "RemoveContainer" containerID="ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.254726 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e"} err="failed to get container status \"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e\": rpc error: code = NotFound desc = could not find container \"ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e\": container with ID starting with ecb34191c7a273d954489c9b09e13fd7fea3cc098044e59c9cc44d5c3c468a8e not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.254748 4744 scope.go:117] "RemoveContainer" containerID="9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.258641 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c"} err="failed to get container status \"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c\": rpc error: code = NotFound desc = could not find container \"9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c\": container with ID starting with 9e79923524d363022615cfb4703512b09ca89b3f6388357073944710c7a1484c not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.258669 4744 scope.go:117] "RemoveContainer" containerID="e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.259511 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-run-httpd\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.259565 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddl6v\" (UniqueName: \"kubernetes.io/projected/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-kube-api-access-ddl6v\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.259604 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-log-httpd\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.259649 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.259693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-scripts\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.259755 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-config-data\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.259784 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.266270 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9"} err="failed to get container status \"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9\": rpc error: code = NotFound desc = could not find container \"e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9\": container with ID starting with e7b07c087248f7d73d621d46f1351e57e41d11094293cab13ede3ce463be02e9 not found: ID does not exist" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.305029 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8892691-3762-4ddf-a9b4-bcf9417acd02" path="/var/lib/kubelet/pods/b8892691-3762-4ddf-a9b4-bcf9417acd02/volumes" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.361331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-run-httpd\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.361604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddl6v\" (UniqueName: \"kubernetes.io/projected/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-kube-api-access-ddl6v\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.361702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-log-httpd\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.361803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.361903 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-scripts\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.361976 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-run-httpd\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.362063 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-config-data\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.362137 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.362150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-log-httpd\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.366222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.366858 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-config-data\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.368239 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-scripts\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.369180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.381373 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddl6v\" (UniqueName: \"kubernetes.io/projected/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-kube-api-access-ddl6v\") pod \"ceilometer-0\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.567111 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.927172 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" event={"ID":"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb","Type":"ContainerStarted","Data":"ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8"} Dec 01 08:36:26 crc kubenswrapper[4744]: I1201 08:36:26.927466 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:27 crc kubenswrapper[4744]: I1201 08:36:27.021978 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" podStartSLOduration=4.021955724 podStartE2EDuration="4.021955724s" podCreationTimestamp="2025-12-01 08:36:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:26.954679178 +0000 UTC m=+1118.943737119" watchObservedRunningTime="2025-12-01 08:36:27.021955724 +0000 UTC m=+1119.011013645" Dec 01 08:36:27 crc kubenswrapper[4744]: I1201 08:36:27.033070 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:27 crc kubenswrapper[4744]: I1201 08:36:27.556144 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 08:36:27 crc kubenswrapper[4744]: I1201 08:36:27.941757 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerStarted","Data":"1f0b46aa74b38721efe15b1113ada0bdd1f29dca55261e486329317628d21fdd"} Dec 01 08:36:28 crc kubenswrapper[4744]: I1201 08:36:28.047980 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:28 crc kubenswrapper[4744]: I1201 08:36:28.953509 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerStarted","Data":"279ddd346f8d6885a926bdd13cd2f10033ff65d0a75a9032142efb1674a25691"} Dec 01 08:36:28 crc kubenswrapper[4744]: I1201 08:36:28.954818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d5b458c99-8sfkr" event={"ID":"3b626810-3162-42db-83b2-9594a48dd935","Type":"ContainerStarted","Data":"54003913e90aa5257f4e0a89701c76c04c5a5b382ffab5ef9608c022b3e14848"} Dec 01 08:36:28 crc kubenswrapper[4744]: I1201 08:36:28.955198 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:28 crc kubenswrapper[4744]: I1201 08:36:28.982013 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5d5b458c99-8sfkr" podStartSLOduration=3.228705067 podStartE2EDuration="5.98199605s" podCreationTimestamp="2025-12-01 08:36:23 +0000 UTC" firstStartedPulling="2025-12-01 08:36:25.045720345 +0000 UTC m=+1117.034778266" lastFinishedPulling="2025-12-01 08:36:27.799011328 +0000 UTC m=+1119.788069249" observedRunningTime="2025-12-01 08:36:28.978757038 +0000 UTC m=+1120.967814959" watchObservedRunningTime="2025-12-01 08:36:28.98199605 +0000 UTC m=+1120.971053971" Dec 01 08:36:29 crc kubenswrapper[4744]: I1201 08:36:29.964129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerStarted","Data":"91d58797246780c729e3e1fdbfc9cc431f2165c5a32c748091f00c51e63af59c"} Dec 01 08:36:29 crc kubenswrapper[4744]: I1201 08:36:29.966423 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" event={"ID":"33b049d4-4934-46bd-8dff-cf9995702c8d","Type":"ContainerStarted","Data":"1e6bf2a3eba3b3c95dfd07efc8e5c4b33126732bfc27bec9d4eab7f87125a469"} Dec 01 08:36:29 crc kubenswrapper[4744]: I1201 08:36:29.986083 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" podStartSLOduration=2.951787223 podStartE2EDuration="6.986066798s" podCreationTimestamp="2025-12-01 08:36:23 +0000 UTC" firstStartedPulling="2025-12-01 08:36:25.179615338 +0000 UTC m=+1117.168673259" lastFinishedPulling="2025-12-01 08:36:29.213894913 +0000 UTC m=+1121.202952834" observedRunningTime="2025-12-01 08:36:29.98119177 +0000 UTC m=+1121.970249701" watchObservedRunningTime="2025-12-01 08:36:29.986066798 +0000 UTC m=+1121.975124719" Dec 01 08:36:30 crc kubenswrapper[4744]: I1201 08:36:30.978559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerStarted","Data":"4dfcdbfda44f684944eebb6bc027f16c1e974a30ab3daea1f18d5893d5a8dbca"} Dec 01 08:36:30 crc kubenswrapper[4744]: I1201 08:36:30.978878 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:31 crc kubenswrapper[4744]: I1201 08:36:31.989207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerStarted","Data":"7e7449ab618120d846da616224c4a99796132a3829085f348a0e11ad06753fb2"} Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.013972 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.591671296 podStartE2EDuration="6.013955017s" podCreationTimestamp="2025-12-01 08:36:26 +0000 UTC" firstStartedPulling="2025-12-01 08:36:27.051814354 +0000 UTC m=+1119.040872275" lastFinishedPulling="2025-12-01 08:36:31.474098055 +0000 UTC m=+1123.463155996" observedRunningTime="2025-12-01 08:36:32.01093433 +0000 UTC m=+1123.999992271" watchObservedRunningTime="2025-12-01 08:36:32.013955017 +0000 UTC m=+1124.003012938" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.706825 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5bf9f8b584-7k2r7"] Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.707862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.723885 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bf9f8b584-7k2r7"] Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.754295 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-56757fd9c7-xswjh"] Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.757777 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.777398 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-f844f6c75-jp9pb"] Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.779330 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.792119 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-56757fd9c7-xswjh"] Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.831328 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f844f6c75-jp9pb"] Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.844450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845016 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx2m5\" (UniqueName: \"kubernetes.io/projected/0dc1a384-3224-46e5-aadc-0421bb72626a-kube-api-access-kx2m5\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data-custom\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845085 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-combined-ca-bundle\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845116 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-config-data\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht8pb\" (UniqueName: \"kubernetes.io/projected/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-kube-api-access-ht8pb\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845260 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845298 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnq88\" (UniqueName: \"kubernetes.io/projected/11429dc7-0531-4503-b66e-f6ecef318ec0-kube-api-access-jnq88\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845318 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-combined-ca-bundle\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data-custom\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845467 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-combined-ca-bundle\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.845506 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-config-data-custom\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947163 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnq88\" (UniqueName: \"kubernetes.io/projected/11429dc7-0531-4503-b66e-f6ecef318ec0-kube-api-access-jnq88\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-combined-ca-bundle\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947238 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data-custom\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947601 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-combined-ca-bundle\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-config-data-custom\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947674 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947863 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx2m5\" (UniqueName: \"kubernetes.io/projected/0dc1a384-3224-46e5-aadc-0421bb72626a-kube-api-access-kx2m5\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data-custom\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-combined-ca-bundle\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-config-data\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.947962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht8pb\" (UniqueName: \"kubernetes.io/projected/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-kube-api-access-ht8pb\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.957696 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-config-data-custom\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.959232 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data-custom\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.960063 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-config-data\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.962764 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-combined-ca-bundle\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.963464 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.964455 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.965136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11429dc7-0531-4503-b66e-f6ecef318ec0-combined-ca-bundle\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.966998 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht8pb\" (UniqueName: \"kubernetes.io/projected/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-kube-api-access-ht8pb\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.970526 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnq88\" (UniqueName: \"kubernetes.io/projected/11429dc7-0531-4503-b66e-f6ecef318ec0-kube-api-access-jnq88\") pod \"heat-engine-5bf9f8b584-7k2r7\" (UID: \"11429dc7-0531-4503-b66e-f6ecef318ec0\") " pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.971090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data-custom\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.971382 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx2m5\" (UniqueName: \"kubernetes.io/projected/0dc1a384-3224-46e5-aadc-0421bb72626a-kube-api-access-kx2m5\") pod \"heat-api-56757fd9c7-xswjh\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.972699 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-combined-ca-bundle\") pod \"heat-cfnapi-f844f6c75-jp9pb\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:32 crc kubenswrapper[4744]: I1201 08:36:32.999703 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:36:33 crc kubenswrapper[4744]: I1201 08:36:33.032270 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:33 crc kubenswrapper[4744]: I1201 08:36:33.047508 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-696c89c747-77nz8" Dec 01 08:36:33 crc kubenswrapper[4744]: I1201 08:36:33.075208 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:33 crc kubenswrapper[4744]: I1201 08:36:33.101867 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.174992 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5d5b458c99-8sfkr"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.175365 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5d5b458c99-8sfkr" podUID="3b626810-3162-42db-83b2-9594a48dd935" containerName="heat-api" containerID="cri-o://54003913e90aa5257f4e0a89701c76c04c5a5b382ffab5ef9608c022b3e14848" gracePeriod=60 Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.198245 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5d5b458c99-8sfkr" podUID="3b626810-3162-42db-83b2-9594a48dd935" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.171:8004/healthcheck\": EOF" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.224455 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5fv45"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.225601 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.233663 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5fv45"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.245452 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5c9d7d48-5qkmn"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.246617 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.249787 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.250016 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.256201 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-869bdb8dbc-7n726"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.256368 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerName="heat-cfnapi" containerID="cri-o://1e6bf2a3eba3b3c95dfd07efc8e5c4b33126732bfc27bec9d4eab7f87125a469" gracePeriod=60 Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.274237 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5c9d7d48-5qkmn"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.278442 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.172:8000/healthcheck\": EOF" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.309730 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.316562 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-68d7bb6659-xxvcz"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.317679 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.321686 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.321873 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.325250 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-k46jn"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.326365 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.336244 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f431-account-create-update-lvs6w"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.338870 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.348599 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.368715 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-k46jn"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373167 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-combined-ca-bundle\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-config-data\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373277 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-internal-tls-certs\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373310 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-config-data-custom\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-combined-ca-bundle\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-public-tls-certs\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-public-tls-certs\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373427 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-config-data\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373472 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbjqr\" (UniqueName: \"kubernetes.io/projected/6abf6646-56a2-4700-9a37-9474036b8865-kube-api-access-nbjqr\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373495 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-internal-tls-certs\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373513 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/606662e8-bd00-4a68-ba9f-26b9d93d175b-operator-scripts\") pod \"nova-api-db-create-5fv45\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373542 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpwm2\" (UniqueName: \"kubernetes.io/projected/a60bb987-1492-435c-9f53-17c3820ce1a0-kube-api-access-zpwm2\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcrgj\" (UniqueName: \"kubernetes.io/projected/606662e8-bd00-4a68-ba9f-26b9d93d175b-kube-api-access-hcrgj\") pod \"nova-api-db-create-5fv45\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.373651 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-config-data-custom\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.423188 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f431-account-create-update-lvs6w"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.448168 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68d7bb6659-xxvcz"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475223 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-internal-tls-certs\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-config-data-custom\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-combined-ca-bundle\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-public-tls-certs\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-public-tls-certs\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475365 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-config-data\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475387 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbjqr\" (UniqueName: \"kubernetes.io/projected/6abf6646-56a2-4700-9a37-9474036b8865-kube-api-access-nbjqr\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475423 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-internal-tls-certs\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475448 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/606662e8-bd00-4a68-ba9f-26b9d93d175b-operator-scripts\") pod \"nova-api-db-create-5fv45\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475478 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5zwg\" (UniqueName: \"kubernetes.io/projected/7230ec92-1f41-491b-a2b5-23aa29cf4d27-kube-api-access-n5zwg\") pod \"nova-api-f431-account-create-update-lvs6w\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475510 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpwm2\" (UniqueName: \"kubernetes.io/projected/a60bb987-1492-435c-9f53-17c3820ce1a0-kube-api-access-zpwm2\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475532 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcrgj\" (UniqueName: \"kubernetes.io/projected/606662e8-bd00-4a68-ba9f-26b9d93d175b-kube-api-access-hcrgj\") pod \"nova-api-db-create-5fv45\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475575 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldq6d\" (UniqueName: \"kubernetes.io/projected/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-kube-api-access-ldq6d\") pod \"nova-cell0-db-create-k46jn\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-config-data-custom\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7230ec92-1f41-491b-a2b5-23aa29cf4d27-operator-scripts\") pod \"nova-api-f431-account-create-update-lvs6w\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475672 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-combined-ca-bundle\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-operator-scripts\") pod \"nova-cell0-db-create-k46jn\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.475715 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-config-data\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.477163 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/606662e8-bd00-4a68-ba9f-26b9d93d175b-operator-scripts\") pod \"nova-api-db-create-5fv45\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.486971 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-internal-tls-certs\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.488210 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-combined-ca-bundle\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.488774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-internal-tls-certs\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.494266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-public-tls-certs\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.494352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-config-data\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.498477 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-sslnq"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.499642 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.502392 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-public-tls-certs\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.503039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-combined-ca-bundle\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.503460 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-config-data\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.505192 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6abf6646-56a2-4700-9a37-9474036b8865-config-data-custom\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.507744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbjqr\" (UniqueName: \"kubernetes.io/projected/6abf6646-56a2-4700-9a37-9474036b8865-kube-api-access-nbjqr\") pod \"heat-cfnapi-68d7bb6659-xxvcz\" (UID: \"6abf6646-56a2-4700-9a37-9474036b8865\") " pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.507702 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcrgj\" (UniqueName: \"kubernetes.io/projected/606662e8-bd00-4a68-ba9f-26b9d93d175b-kube-api-access-hcrgj\") pod \"nova-api-db-create-5fv45\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.505434 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpwm2\" (UniqueName: \"kubernetes.io/projected/a60bb987-1492-435c-9f53-17c3820ce1a0-kube-api-access-zpwm2\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.510277 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a60bb987-1492-435c-9f53-17c3820ce1a0-config-data-custom\") pod \"heat-api-5c9d7d48-5qkmn\" (UID: \"a60bb987-1492-435c-9f53-17c3820ce1a0\") " pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.520128 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-sslnq"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.552491 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-5hrgd"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.553086 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerName="dnsmasq-dns" containerID="cri-o://77e44ebd4d9ce670b2487d0b6fa54296c4d22e5d154f182a92bf21c1af24f22e" gracePeriod=10 Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.564680 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.586362 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8cec-account-create-update-xlrzh"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.587467 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.588437 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.588648 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7230ec92-1f41-491b-a2b5-23aa29cf4d27-operator-scripts\") pod \"nova-api-f431-account-create-update-lvs6w\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.588682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-operator-scripts\") pod \"nova-cell0-db-create-k46jn\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.588717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5172715-95f9-48a5-bf83-8fbe12093298-operator-scripts\") pod \"nova-cell1-db-create-sslnq\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.588762 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75s66\" (UniqueName: \"kubernetes.io/projected/e5172715-95f9-48a5-bf83-8fbe12093298-kube-api-access-75s66\") pod \"nova-cell1-db-create-sslnq\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.588808 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5zwg\" (UniqueName: \"kubernetes.io/projected/7230ec92-1f41-491b-a2b5-23aa29cf4d27-kube-api-access-n5zwg\") pod \"nova-api-f431-account-create-update-lvs6w\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.588858 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldq6d\" (UniqueName: \"kubernetes.io/projected/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-kube-api-access-ldq6d\") pod \"nova-cell0-db-create-k46jn\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.589041 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.589861 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-operator-scripts\") pod \"nova-cell0-db-create-k46jn\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.590637 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7230ec92-1f41-491b-a2b5-23aa29cf4d27-operator-scripts\") pod \"nova-api-f431-account-create-update-lvs6w\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.599124 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8cec-account-create-update-xlrzh"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.609641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5zwg\" (UniqueName: \"kubernetes.io/projected/7230ec92-1f41-491b-a2b5-23aa29cf4d27-kube-api-access-n5zwg\") pod \"nova-api-f431-account-create-update-lvs6w\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.629067 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldq6d\" (UniqueName: \"kubernetes.io/projected/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-kube-api-access-ldq6d\") pod \"nova-cell0-db-create-k46jn\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.646834 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.663229 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-fe16-account-create-update-h8bdl"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.665600 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.667524 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.667967 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.686590 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-fe16-account-create-update-h8bdl"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.690168 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59632a6c-e03f-45dd-a08a-561603cb97b5-operator-scripts\") pod \"nova-cell0-8cec-account-create-update-xlrzh\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.690234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5172715-95f9-48a5-bf83-8fbe12093298-operator-scripts\") pod \"nova-cell1-db-create-sslnq\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.690285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75s66\" (UniqueName: \"kubernetes.io/projected/e5172715-95f9-48a5-bf83-8fbe12093298-kube-api-access-75s66\") pod \"nova-cell1-db-create-sslnq\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.690335 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7s4m\" (UniqueName: \"kubernetes.io/projected/59632a6c-e03f-45dd-a08a-561603cb97b5-kube-api-access-x7s4m\") pod \"nova-cell0-8cec-account-create-update-xlrzh\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.690648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.691964 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5172715-95f9-48a5-bf83-8fbe12093298-operator-scripts\") pod \"nova-cell1-db-create-sslnq\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.714247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75s66\" (UniqueName: \"kubernetes.io/projected/e5172715-95f9-48a5-bf83-8fbe12093298-kube-api-access-75s66\") pod \"nova-cell1-db-create-sslnq\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.792846 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7s4m\" (UniqueName: \"kubernetes.io/projected/59632a6c-e03f-45dd-a08a-561603cb97b5-kube-api-access-x7s4m\") pod \"nova-cell0-8cec-account-create-update-xlrzh\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.792932 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjdmw\" (UniqueName: \"kubernetes.io/projected/65092f5c-34d0-4eaa-9f53-dea23cc222d4-kube-api-access-wjdmw\") pod \"nova-cell1-fe16-account-create-update-h8bdl\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.792998 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59632a6c-e03f-45dd-a08a-561603cb97b5-operator-scripts\") pod \"nova-cell0-8cec-account-create-update-xlrzh\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.793069 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65092f5c-34d0-4eaa-9f53-dea23cc222d4-operator-scripts\") pod \"nova-cell1-fe16-account-create-update-h8bdl\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.795602 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59632a6c-e03f-45dd-a08a-561603cb97b5-operator-scripts\") pod \"nova-cell0-8cec-account-create-update-xlrzh\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.811088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7s4m\" (UniqueName: \"kubernetes.io/projected/59632a6c-e03f-45dd-a08a-561603cb97b5-kube-api-access-x7s4m\") pod \"nova-cell0-8cec-account-create-update-xlrzh\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.899302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjdmw\" (UniqueName: \"kubernetes.io/projected/65092f5c-34d0-4eaa-9f53-dea23cc222d4-kube-api-access-wjdmw\") pod \"nova-cell1-fe16-account-create-update-h8bdl\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.899449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65092f5c-34d0-4eaa-9f53-dea23cc222d4-operator-scripts\") pod \"nova-cell1-fe16-account-create-update-h8bdl\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.900111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65092f5c-34d0-4eaa-9f53-dea23cc222d4-operator-scripts\") pod \"nova-cell1-fe16-account-create-update-h8bdl\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.910755 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:34 crc kubenswrapper[4744]: I1201 08:36:34.920136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjdmw\" (UniqueName: \"kubernetes.io/projected/65092f5c-34d0-4eaa-9f53-dea23cc222d4-kube-api-access-wjdmw\") pod \"nova-cell1-fe16-account-create-update-h8bdl\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.015012 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.017888 4744 generic.go:334] "Generic (PLEG): container finished" podID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerID="77e44ebd4d9ce670b2487d0b6fa54296c4d22e5d154f182a92bf21c1af24f22e" exitCode=0 Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.018121 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-central-agent" containerID="cri-o://279ddd346f8d6885a926bdd13cd2f10033ff65d0a75a9032142efb1674a25691" gracePeriod=30 Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.018382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" event={"ID":"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64","Type":"ContainerDied","Data":"77e44ebd4d9ce670b2487d0b6fa54296c4d22e5d154f182a92bf21c1af24f22e"} Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.018676 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="proxy-httpd" containerID="cri-o://7e7449ab618120d846da616224c4a99796132a3829085f348a0e11ad06753fb2" gracePeriod=30 Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.018730 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="sg-core" containerID="cri-o://4dfcdbfda44f684944eebb6bc027f16c1e974a30ab3daea1f18d5893d5a8dbca" gracePeriod=30 Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.018763 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-notification-agent" containerID="cri-o://91d58797246780c729e3e1fdbfc9cc431f2165c5a32c748091f00c51e63af59c" gracePeriod=30 Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.071163 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.082987 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.690916 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.691514 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-log" containerID="cri-o://fa8f79b5a76b19113710fe7c2f739a769d63613b3da2f9df83d8cc48399ddd90" gracePeriod=30 Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.691604 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-httpd" containerID="cri-o://e6cd242e2347cb88215c4090a3ce52c86d093b6f7e074e034a268a8206eb0394" gracePeriod=30 Dec 01 08:36:35 crc kubenswrapper[4744]: I1201 08:36:35.926021 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.161:5353: connect: connection refused" Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.034999 4744 generic.go:334] "Generic (PLEG): container finished" podID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerID="7e7449ab618120d846da616224c4a99796132a3829085f348a0e11ad06753fb2" exitCode=0 Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.035034 4744 generic.go:334] "Generic (PLEG): container finished" podID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerID="4dfcdbfda44f684944eebb6bc027f16c1e974a30ab3daea1f18d5893d5a8dbca" exitCode=2 Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.035041 4744 generic.go:334] "Generic (PLEG): container finished" podID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerID="91d58797246780c729e3e1fdbfc9cc431f2165c5a32c748091f00c51e63af59c" exitCode=0 Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.035049 4744 generic.go:334] "Generic (PLEG): container finished" podID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerID="279ddd346f8d6885a926bdd13cd2f10033ff65d0a75a9032142efb1674a25691" exitCode=0 Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.035082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerDied","Data":"7e7449ab618120d846da616224c4a99796132a3829085f348a0e11ad06753fb2"} Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.035106 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerDied","Data":"4dfcdbfda44f684944eebb6bc027f16c1e974a30ab3daea1f18d5893d5a8dbca"} Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.035117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerDied","Data":"91d58797246780c729e3e1fdbfc9cc431f2165c5a32c748091f00c51e63af59c"} Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.035126 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerDied","Data":"279ddd346f8d6885a926bdd13cd2f10033ff65d0a75a9032142efb1674a25691"} Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.040743 4744 generic.go:334] "Generic (PLEG): container finished" podID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerID="fa8f79b5a76b19113710fe7c2f739a769d63613b3da2f9df83d8cc48399ddd90" exitCode=143 Dec 01 08:36:36 crc kubenswrapper[4744]: I1201 08:36:36.040766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1691c44a-f2ba-43eb-9ff3-a6a25b352d78","Type":"ContainerDied","Data":"fa8f79b5a76b19113710fe7c2f739a769d63613b3da2f9df83d8cc48399ddd90"} Dec 01 08:36:38 crc kubenswrapper[4744]: I1201 08:36:38.823257 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:38 crc kubenswrapper[4744]: I1201 08:36:38.991960 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-svc\") pod \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " Dec 01 08:36:38 crc kubenswrapper[4744]: I1201 08:36:38.992039 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-config\") pod \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " Dec 01 08:36:38 crc kubenswrapper[4744]: I1201 08:36:38.992094 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-sb\") pod \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " Dec 01 08:36:38 crc kubenswrapper[4744]: I1201 08:36:38.992195 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smwl9\" (UniqueName: \"kubernetes.io/projected/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-kube-api-access-smwl9\") pod \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " Dec 01 08:36:38 crc kubenswrapper[4744]: I1201 08:36:38.992231 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-swift-storage-0\") pod \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " Dec 01 08:36:38 crc kubenswrapper[4744]: I1201 08:36:38.992287 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-nb\") pod \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\" (UID: \"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.012448 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-kube-api-access-smwl9" (OuterVolumeSpecName: "kube-api-access-smwl9") pod "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" (UID: "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64"). InnerVolumeSpecName "kube-api-access-smwl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.065089 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" (UID: "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.065185 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-config" (OuterVolumeSpecName: "config") pod "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" (UID: "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.069778 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" (UID: "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.077462 4744 generic.go:334] "Generic (PLEG): container finished" podID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerID="e6cd242e2347cb88215c4090a3ce52c86d093b6f7e074e034a268a8206eb0394" exitCode=0 Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.077526 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1691c44a-f2ba-43eb-9ff3-a6a25b352d78","Type":"ContainerDied","Data":"e6cd242e2347cb88215c4090a3ce52c86d093b6f7e074e034a268a8206eb0394"} Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.086306 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" (UID: "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.087522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"19db2e11-d46c-4fec-a7a4-b3afe22af5fa","Type":"ContainerDied","Data":"1f0b46aa74b38721efe15b1113ada0bdd1f29dca55261e486329317628d21fdd"} Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.087557 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f0b46aa74b38721efe15b1113ada0bdd1f29dca55261e486329317628d21fdd" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.088197 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.091559 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" (UID: "87a3a2b6-e743-46dd-bcdb-a6f598ac8c64"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.092298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" event={"ID":"87a3a2b6-e743-46dd-bcdb-a6f598ac8c64","Type":"ContainerDied","Data":"dab3eedac35259b7bb90eed5949e5f4ffc756690766947672bec32b12d368831"} Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.092333 4744 scope.go:117] "RemoveContainer" containerID="77e44ebd4d9ce670b2487d0b6fa54296c4d22e5d154f182a92bf21c1af24f22e" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.092482 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-5hrgd" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.103734 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smwl9\" (UniqueName: \"kubernetes.io/projected/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-kube-api-access-smwl9\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.103762 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.103771 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.103781 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.103793 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.103807 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.125728 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"181b6931-2498-4979-a9b1-f6cbc925c78e","Type":"ContainerStarted","Data":"ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b"} Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.167651 4744 scope.go:117] "RemoveContainer" containerID="3d9f0daedf8f3b1b963778314cf5092b15f5b402e3e059242fa9404b244dc83b" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.205857 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-run-httpd\") pod \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.205897 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-scripts\") pod \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.205906 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.336235017 podStartE2EDuration="20.205881526s" podCreationTimestamp="2025-12-01 08:36:19 +0000 UTC" firstStartedPulling="2025-12-01 08:36:20.667917229 +0000 UTC m=+1112.656975150" lastFinishedPulling="2025-12-01 08:36:38.537563738 +0000 UTC m=+1130.526621659" observedRunningTime="2025-12-01 08:36:39.159948138 +0000 UTC m=+1131.149006059" watchObservedRunningTime="2025-12-01 08:36:39.205881526 +0000 UTC m=+1131.194939447" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.205990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-config-data\") pod \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.206015 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-combined-ca-bundle\") pod \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.206036 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-sg-core-conf-yaml\") pod \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.206156 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddl6v\" (UniqueName: \"kubernetes.io/projected/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-kube-api-access-ddl6v\") pod \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.206228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-log-httpd\") pod \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\" (UID: \"19db2e11-d46c-4fec-a7a4-b3afe22af5fa\") " Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.206985 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "19db2e11-d46c-4fec-a7a4-b3afe22af5fa" (UID: "19db2e11-d46c-4fec-a7a4-b3afe22af5fa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.207097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "19db2e11-d46c-4fec-a7a4-b3afe22af5fa" (UID: "19db2e11-d46c-4fec-a7a4-b3afe22af5fa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.208558 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-5hrgd"] Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.220664 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-scripts" (OuterVolumeSpecName: "scripts") pod "19db2e11-d46c-4fec-a7a4-b3afe22af5fa" (UID: "19db2e11-d46c-4fec-a7a4-b3afe22af5fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.220705 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-kube-api-access-ddl6v" (OuterVolumeSpecName: "kube-api-access-ddl6v") pod "19db2e11-d46c-4fec-a7a4-b3afe22af5fa" (UID: "19db2e11-d46c-4fec-a7a4-b3afe22af5fa"). InnerVolumeSpecName "kube-api-access-ddl6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.235177 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-5hrgd"] Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.240476 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "19db2e11-d46c-4fec-a7a4-b3afe22af5fa" (UID: "19db2e11-d46c-4fec-a7a4-b3afe22af5fa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.296355 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19db2e11-d46c-4fec-a7a4-b3afe22af5fa" (UID: "19db2e11-d46c-4fec-a7a4-b3afe22af5fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.311201 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.312462 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.312539 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.312607 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.312690 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.312744 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddl6v\" (UniqueName: \"kubernetes.io/projected/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-kube-api-access-ddl6v\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.346258 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-config-data" (OuterVolumeSpecName: "config-data") pod "19db2e11-d46c-4fec-a7a4-b3afe22af5fa" (UID: "19db2e11-d46c-4fec-a7a4-b3afe22af5fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.414533 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19db2e11-d46c-4fec-a7a4-b3afe22af5fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.659549 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5d5b458c99-8sfkr" podUID="3b626810-3162-42db-83b2-9594a48dd935" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.171:8004/healthcheck\": read tcp 10.217.0.2:40124->10.217.0.171:8004: read: connection reset by peer" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.659967 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5d5b458c99-8sfkr" podUID="3b626810-3162-42db-83b2-9594a48dd935" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.171:8004/healthcheck\": dial tcp 10.217.0.171:8004: connect: connection refused" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.676558 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.172:8000/healthcheck\": read tcp 10.217.0.2:34044->10.217.0.172:8000: read: connection reset by peer" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.677038 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.172:8000/healthcheck\": dial tcp 10.217.0.172:8000: connect: connection refused" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.951562 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.964157 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-fe16-account-create-update-h8bdl"] Dec 01 08:36:39 crc kubenswrapper[4744]: I1201 08:36:39.995525 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8cec-account-create-update-xlrzh"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.011441 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bf9f8b584-7k2r7"] Dec 01 08:36:40 crc kubenswrapper[4744]: W1201 08:36:40.046086 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda60bb987_1492_435c_9f53_17c3820ce1a0.slice/crio-690bf8d624da6dff1fb16ede3e459cdfc99e92c97567c4506dd2b28172627c53 WatchSource:0}: Error finding container 690bf8d624da6dff1fb16ede3e459cdfc99e92c97567c4506dd2b28172627c53: Status 404 returned error can't find the container with id 690bf8d624da6dff1fb16ede3e459cdfc99e92c97567c4506dd2b28172627c53 Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.060386 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-sslnq"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.076686 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-k46jn"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.084993 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f844f6c75-jp9pb"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.096202 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68d7bb6659-xxvcz"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.105648 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5c9d7d48-5qkmn"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.112800 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5fv45"] Dec 01 08:36:40 crc kubenswrapper[4744]: W1201 08:36:40.132944 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7230ec92_1f41_491b_a2b5_23aa29cf4d27.slice/crio-77c075234dc60dfd106f57d60f4cb94174a2a77e6091efaae1bb8936c9501919 WatchSource:0}: Error finding container 77c075234dc60dfd106f57d60f4cb94174a2a77e6091efaae1bb8936c9501919: Status 404 returned error can't find the container with id 77c075234dc60dfd106f57d60f4cb94174a2a77e6091efaae1bb8936c9501919 Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.135895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-logs\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.136501 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-logs" (OuterVolumeSpecName: "logs") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.136552 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-combined-ca-bundle\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.136601 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-httpd-run\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.136936 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-public-tls-certs\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.136981 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-scripts\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.137100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-config-data\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.137192 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntnbh\" (UniqueName: \"kubernetes.io/projected/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-kube-api-access-ntnbh\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.137478 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\" (UID: \"1691c44a-f2ba-43eb-9ff3-a6a25b352d78\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.137650 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.138381 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.143706 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-scripts" (OuterVolumeSpecName: "scripts") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.143797 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.146216 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f431-account-create-update-lvs6w"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.148134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sslnq" event={"ID":"e5172715-95f9-48a5-bf83-8fbe12093298","Type":"ContainerStarted","Data":"3b95f9a8808eb75db90c4e288399856c08c3060f15d3c6f42fb8623931b65182"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.148213 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-kube-api-access-ntnbh" (OuterVolumeSpecName: "kube-api-access-ntnbh") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "kube-api-access-ntnbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.156134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bf9f8b584-7k2r7" event={"ID":"11429dc7-0531-4503-b66e-f6ecef318ec0","Type":"ContainerStarted","Data":"552a753427d18bc8337aae3ad6e746693bc1c4e87c6533001f8ee59b0e76b65e"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.156676 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-56757fd9c7-xswjh"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.166772 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" event={"ID":"9223dbb6-40f1-4946-b5e3-e84a1e25aa26","Type":"ContainerStarted","Data":"ead31e7842ab61d4413f0f673306fb4b253db0a8b28e5c6e19b70a3a02683df8"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.169072 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.169326 4744 generic.go:334] "Generic (PLEG): container finished" podID="3b626810-3162-42db-83b2-9594a48dd935" containerID="54003913e90aa5257f4e0a89701c76c04c5a5b382ffab5ef9608c022b3e14848" exitCode=0 Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.169487 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d5b458c99-8sfkr" event={"ID":"3b626810-3162-42db-83b2-9594a48dd935","Type":"ContainerDied","Data":"54003913e90aa5257f4e0a89701c76c04c5a5b382ffab5ef9608c022b3e14848"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.170689 4744 scope.go:117] "RemoveContainer" containerID="54003913e90aa5257f4e0a89701c76c04c5a5b382ffab5ef9608c022b3e14848" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.172518 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5fv45" event={"ID":"606662e8-bd00-4a68-ba9f-26b9d93d175b","Type":"ContainerStarted","Data":"80595464f83be08d3bc42ceabc93e31d0c229295156fa6a45c8bfedd55791212"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.174140 4744 generic.go:334] "Generic (PLEG): container finished" podID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerID="1e6bf2a3eba3b3c95dfd07efc8e5c4b33126732bfc27bec9d4eab7f87125a469" exitCode=0 Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.174193 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" event={"ID":"33b049d4-4934-46bd-8dff-cf9995702c8d","Type":"ContainerDied","Data":"1e6bf2a3eba3b3c95dfd07efc8e5c4b33126732bfc27bec9d4eab7f87125a469"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.175683 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c9d7d48-5qkmn" event={"ID":"a60bb987-1492-435c-9f53-17c3820ce1a0","Type":"ContainerStarted","Data":"690bf8d624da6dff1fb16ede3e459cdfc99e92c97567c4506dd2b28172627c53"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.176458 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" event={"ID":"65092f5c-34d0-4eaa-9f53-dea23cc222d4","Type":"ContainerStarted","Data":"dafdc58a87e0394a1efcba0b6638f52e16c412bd723648368d793eca89ae7a7b"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.198618 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" event={"ID":"6abf6646-56a2-4700-9a37-9474036b8865","Type":"ContainerStarted","Data":"b17c55c5dde4bc7bd6c16e0a7bafa96cf1efa385af4afd9e8d309aa456bfca18"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.205280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-k46jn" event={"ID":"80d42461-4c7a-4e38-8ac7-1261d5fcb37a","Type":"ContainerStarted","Data":"fc72d10a9e48af520b0ce8ddefa472e2588f0301742a0c85026618752f9f2b21"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.211374 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1691c44a-f2ba-43eb-9ff3-a6a25b352d78","Type":"ContainerDied","Data":"0d33f494f418f0023078563f0809b0ab21d55b14b3f06995bc9ba03d989746c5"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.211480 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.220140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" event={"ID":"59632a6c-e03f-45dd-a08a-561603cb97b5","Type":"ContainerStarted","Data":"8d2a7d55cbedc275d40c187a8ecba14645188463665af7caa30638aae9ce853f"} Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.220242 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.242328 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.242362 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.242372 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntnbh\" (UniqueName: \"kubernetes.io/projected/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-kube-api-access-ntnbh\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.242397 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.269450 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.269542 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.322284 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" path="/var/lib/kubelet/pods/87a3a2b6-e743-46dd-bcdb-a6f598ac8c64/volumes" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.343636 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76gd7\" (UniqueName: \"kubernetes.io/projected/3b626810-3162-42db-83b2-9594a48dd935-kube-api-access-76gd7\") pod \"3b626810-3162-42db-83b2-9594a48dd935\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.345480 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data\") pod \"3b626810-3162-42db-83b2-9594a48dd935\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.345599 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-combined-ca-bundle\") pod \"3b626810-3162-42db-83b2-9594a48dd935\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.345739 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data-custom\") pod \"3b626810-3162-42db-83b2-9594a48dd935\" (UID: \"3b626810-3162-42db-83b2-9594a48dd935\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.346471 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.362833 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b626810-3162-42db-83b2-9594a48dd935-kube-api-access-76gd7" (OuterVolumeSpecName: "kube-api-access-76gd7") pod "3b626810-3162-42db-83b2-9594a48dd935" (UID: "3b626810-3162-42db-83b2-9594a48dd935"). InnerVolumeSpecName "kube-api-access-76gd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.364426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3b626810-3162-42db-83b2-9594a48dd935" (UID: "3b626810-3162-42db-83b2-9594a48dd935"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.376626 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.384503 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.406736 4744 scope.go:117] "RemoveContainer" containerID="e6cd242e2347cb88215c4090a3ce52c86d093b6f7e074e034a268a8206eb0394" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.442834 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b626810-3162-42db-83b2-9594a48dd935" (UID: "3b626810-3162-42db-83b2-9594a48dd935"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.446995 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-config-data" (OuterVolumeSpecName: "config-data") pod "1691c44a-f2ba-43eb-9ff3-a6a25b352d78" (UID: "1691c44a-f2ba-43eb-9ff3-a6a25b352d78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.448211 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76gd7\" (UniqueName: \"kubernetes.io/projected/3b626810-3162-42db-83b2-9594a48dd935-kube-api-access-76gd7\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.448319 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.448395 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.449003 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.449070 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.449125 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1691c44a-f2ba-43eb-9ff3-a6a25b352d78-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.541308 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data" (OuterVolumeSpecName: "config-data") pod "3b626810-3162-42db-83b2-9594a48dd935" (UID: "3b626810-3162-42db-83b2-9594a48dd935"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.550970 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b626810-3162-42db-83b2-9594a48dd935-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696005 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696060 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696511 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerName="dnsmasq-dns" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696530 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerName="dnsmasq-dns" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696541 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b626810-3162-42db-83b2-9594a48dd935" containerName="heat-api" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696547 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b626810-3162-42db-83b2-9594a48dd935" containerName="heat-api" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696558 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerName="init" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696580 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerName="init" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696592 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-httpd" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696598 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-httpd" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696606 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-central-agent" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696612 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-central-agent" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696625 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="sg-core" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696631 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="sg-core" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696662 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="proxy-httpd" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696669 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="proxy-httpd" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696692 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-notification-agent" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696699 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-notification-agent" Dec 01 08:36:40 crc kubenswrapper[4744]: E1201 08:36:40.696710 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-log" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696718 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-log" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696924 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-httpd" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696938 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="87a3a2b6-e743-46dd-bcdb-a6f598ac8c64" containerName="dnsmasq-dns" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696947 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-notification-agent" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696981 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="proxy-httpd" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.696993 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" containerName="glance-log" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.697006 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="sg-core" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.697018 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" containerName="ceilometer-central-agent" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.697030 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b626810-3162-42db-83b2-9594a48dd935" containerName="heat-api" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.699776 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.699805 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.699984 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.700145 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-log" containerID="cri-o://3e9efe156cc4be4376971c5a296054f22c889d63d7a2cf1180f1f935b8ec5791" gracePeriod=30 Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.700277 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-httpd" containerID="cri-o://2411f961caa796d67d32b9f30533c1102ca0802392405051a9b1df63f83097a2" gracePeriod=30 Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.702437 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.703746 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.725978 4744 scope.go:117] "RemoveContainer" containerID="fa8f79b5a76b19113710fe7c2f739a769d63613b3da2f9df83d8cc48399ddd90" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.749311 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.856802 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7ljh\" (UniqueName: \"kubernetes.io/projected/33b049d4-4934-46bd-8dff-cf9995702c8d-kube-api-access-f7ljh\") pod \"33b049d4-4934-46bd-8dff-cf9995702c8d\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.857238 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-combined-ca-bundle\") pod \"33b049d4-4934-46bd-8dff-cf9995702c8d\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.857326 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data\") pod \"33b049d4-4934-46bd-8dff-cf9995702c8d\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.857511 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data-custom\") pod \"33b049d4-4934-46bd-8dff-cf9995702c8d\" (UID: \"33b049d4-4934-46bd-8dff-cf9995702c8d\") " Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.857954 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zfkm\" (UniqueName: \"kubernetes.io/projected/45050404-4ef1-4acb-9bdf-ca9a821fdae7-kube-api-access-5zfkm\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.858099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.858250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-config-data\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.858353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.858462 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-scripts\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.858625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-log-httpd\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.858719 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-run-httpd\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.861889 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b049d4-4934-46bd-8dff-cf9995702c8d-kube-api-access-f7ljh" (OuterVolumeSpecName: "kube-api-access-f7ljh") pod "33b049d4-4934-46bd-8dff-cf9995702c8d" (UID: "33b049d4-4934-46bd-8dff-cf9995702c8d"). InnerVolumeSpecName "kube-api-access-f7ljh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.862186 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "33b049d4-4934-46bd-8dff-cf9995702c8d" (UID: "33b049d4-4934-46bd-8dff-cf9995702c8d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.882017 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33b049d4-4934-46bd-8dff-cf9995702c8d" (UID: "33b049d4-4934-46bd-8dff-cf9995702c8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.915798 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data" (OuterVolumeSpecName: "config-data") pod "33b049d4-4934-46bd-8dff-cf9995702c8d" (UID: "33b049d4-4934-46bd-8dff-cf9995702c8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.960820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zfkm\" (UniqueName: \"kubernetes.io/projected/45050404-4ef1-4acb-9bdf-ca9a821fdae7-kube-api-access-5zfkm\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.960956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-config-data\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961134 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-scripts\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961207 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-log-httpd\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961253 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-run-httpd\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961752 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961779 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961805 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33b049d4-4934-46bd-8dff-cf9995702c8d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.961838 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7ljh\" (UniqueName: \"kubernetes.io/projected/33b049d4-4934-46bd-8dff-cf9995702c8d-kube-api-access-f7ljh\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.962432 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-log-httpd\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.962972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-run-httpd\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:40 crc kubenswrapper[4744]: I1201 08:36:40.981840 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zfkm\" (UniqueName: \"kubernetes.io/projected/45050404-4ef1-4acb-9bdf-ca9a821fdae7-kube-api-access-5zfkm\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.014478 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.014789 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-scripts\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.016163 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-config-data\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.020920 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " pod="openstack/ceilometer-0" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.233912 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.240040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bf9f8b584-7k2r7" event={"ID":"11429dc7-0531-4503-b66e-f6ecef318ec0","Type":"ContainerStarted","Data":"09ede97f453de3093a47e9f9bf2c0ad9163cb8cea884d140dcd7d941a1ff1f18"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.240314 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.257745 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5fv45" event={"ID":"606662e8-bd00-4a68-ba9f-26b9d93d175b","Type":"ContainerStarted","Data":"78e0eae410b90b6f7bcfb7f22d1b6e9e69f6b122e21b37edf7d6df78fe8e20d1"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.262454 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f431-account-create-update-lvs6w" event={"ID":"7230ec92-1f41-491b-a2b5-23aa29cf4d27","Type":"ContainerStarted","Data":"c5449e1ced9108117cf357cc5cde6963d71935c25dddf4f9c40fdfbab1f0aedb"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.262488 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f431-account-create-update-lvs6w" event={"ID":"7230ec92-1f41-491b-a2b5-23aa29cf4d27","Type":"ContainerStarted","Data":"77c075234dc60dfd106f57d60f4cb94174a2a77e6091efaae1bb8936c9501919"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.264191 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5bf9f8b584-7k2r7" podStartSLOduration=9.264172201 podStartE2EDuration="9.264172201s" podCreationTimestamp="2025-12-01 08:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.258915201 +0000 UTC m=+1133.247973122" watchObservedRunningTime="2025-12-01 08:36:41.264172201 +0000 UTC m=+1133.253230122" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.269075 4744 generic.go:334] "Generic (PLEG): container finished" podID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerID="3e9efe156cc4be4376971c5a296054f22c889d63d7a2cf1180f1f935b8ec5791" exitCode=143 Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.269131 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a0a609f-e643-4e6b-b0f8-428fd792fce9","Type":"ContainerDied","Data":"3e9efe156cc4be4376971c5a296054f22c889d63d7a2cf1180f1f935b8ec5791"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.288865 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c9d7d48-5qkmn" event={"ID":"a60bb987-1492-435c-9f53-17c3820ce1a0","Type":"ContainerStarted","Data":"5b697e92d0afd814b118d204946bd6ff6ca32e1ccb7f2e4023e982a3fd043ffc"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.289960 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.311306 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-f431-account-create-update-lvs6w" podStartSLOduration=7.311291162 podStartE2EDuration="7.311291162s" podCreationTimestamp="2025-12-01 08:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.304513759 +0000 UTC m=+1133.293571680" watchObservedRunningTime="2025-12-01 08:36:41.311291162 +0000 UTC m=+1133.300349073" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.311630 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-5fv45" podStartSLOduration=7.311625892 podStartE2EDuration="7.311625892s" podCreationTimestamp="2025-12-01 08:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.28206395 +0000 UTC m=+1133.271121871" watchObservedRunningTime="2025-12-01 08:36:41.311625892 +0000 UTC m=+1133.300683813" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.318037 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" event={"ID":"6abf6646-56a2-4700-9a37-9474036b8865","Type":"ContainerStarted","Data":"f62e462fcbbf2a573ac395fb6913290e1a7dece4b7ff25fa904eb13dd20e7736"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.318114 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.324037 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5c9d7d48-5qkmn" podStartSLOduration=7.324029165 podStartE2EDuration="7.324029165s" podCreationTimestamp="2025-12-01 08:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.320156285 +0000 UTC m=+1133.309214206" watchObservedRunningTime="2025-12-01 08:36:41.324029165 +0000 UTC m=+1133.313087086" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.335336 4744 generic.go:334] "Generic (PLEG): container finished" podID="80d42461-4c7a-4e38-8ac7-1261d5fcb37a" containerID="c1420364bf8ee1eb684aa6d79a949049e0a5003a9c0df563fb01f07b375bfb9f" exitCode=0 Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.335473 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-k46jn" event={"ID":"80d42461-4c7a-4e38-8ac7-1261d5fcb37a","Type":"ContainerDied","Data":"c1420364bf8ee1eb684aa6d79a949049e0a5003a9c0df563fb01f07b375bfb9f"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.347266 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" podStartSLOduration=7.347242686 podStartE2EDuration="7.347242686s" podCreationTimestamp="2025-12-01 08:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.344008474 +0000 UTC m=+1133.333066395" watchObservedRunningTime="2025-12-01 08:36:41.347242686 +0000 UTC m=+1133.336300607" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.349150 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" event={"ID":"59632a6c-e03f-45dd-a08a-561603cb97b5","Type":"ContainerStarted","Data":"d8419e9a848a43bb0186359d12ee3dcf0e7cd46a464cf05568b1c2eb0ca7556e"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.355455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d5b458c99-8sfkr" event={"ID":"3b626810-3162-42db-83b2-9594a48dd935","Type":"ContainerDied","Data":"57e9d21ed6a1b258b7c46f35706d343bb4a78c23b2ea8b06d09edcaa3b54736a"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.355562 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d5b458c99-8sfkr" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.369848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" event={"ID":"33b049d4-4934-46bd-8dff-cf9995702c8d","Type":"ContainerDied","Data":"4bee75623a79eae1198797b9bc35e43f81d3a71333d02af1e06c13f3d0ae0252"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.369917 4744 scope.go:117] "RemoveContainer" containerID="1e6bf2a3eba3b3c95dfd07efc8e5c4b33126732bfc27bec9d4eab7f87125a469" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.370065 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-869bdb8dbc-7n726" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.389739 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" event={"ID":"9223dbb6-40f1-4946-b5e3-e84a1e25aa26","Type":"ContainerStarted","Data":"16a6e0a17275df1778cff2fd6fc4d9e78c3d3c353e53d816498048caf593d379"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.390679 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.396642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" event={"ID":"65092f5c-34d0-4eaa-9f53-dea23cc222d4","Type":"ContainerStarted","Data":"7f2960ab28ffbc55a4e737a7b3e8ba326b0ddfa06f7cac5faa617e8eb314a7e8"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.408695 4744 generic.go:334] "Generic (PLEG): container finished" podID="e5172715-95f9-48a5-bf83-8fbe12093298" containerID="d66a177ed2e123f2913d76f6d902840e58acedd6aa55d4d73aa798862db4035d" exitCode=0 Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.408795 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sslnq" event={"ID":"e5172715-95f9-48a5-bf83-8fbe12093298","Type":"ContainerDied","Data":"d66a177ed2e123f2913d76f6d902840e58acedd6aa55d4d73aa798862db4035d"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.412774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56757fd9c7-xswjh" event={"ID":"0dc1a384-3224-46e5-aadc-0421bb72626a","Type":"ContainerStarted","Data":"2af749052de498012cd9a145bbaccbdd326ca98145b2315ea17dcbac1f351917"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.412803 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56757fd9c7-xswjh" event={"ID":"0dc1a384-3224-46e5-aadc-0421bb72626a","Type":"ContainerStarted","Data":"6ab0f5f23dac41b15548145cf5129567217b3d10d5f301f43e982d66df7db1c5"} Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.412919 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.456516 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" podStartSLOduration=7.456497717 podStartE2EDuration="7.456497717s" podCreationTimestamp="2025-12-01 08:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.397928349 +0000 UTC m=+1133.386986270" watchObservedRunningTime="2025-12-01 08:36:41.456497717 +0000 UTC m=+1133.445555638" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.470857 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5d5b458c99-8sfkr"] Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.478660 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5d5b458c99-8sfkr"] Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.500039 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" podStartSLOduration=9.500019446 podStartE2EDuration="9.500019446s" podCreationTimestamp="2025-12-01 08:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.430369143 +0000 UTC m=+1133.419427064" watchObservedRunningTime="2025-12-01 08:36:41.500019446 +0000 UTC m=+1133.489077367" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.502067 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" podStartSLOduration=7.502060734 podStartE2EDuration="7.502060734s" podCreationTimestamp="2025-12-01 08:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.45203659 +0000 UTC m=+1133.441094511" watchObservedRunningTime="2025-12-01 08:36:41.502060734 +0000 UTC m=+1133.491118655" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.554757 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-869bdb8dbc-7n726"] Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.563470 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-869bdb8dbc-7n726"] Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.580064 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-56757fd9c7-xswjh" podStartSLOduration=9.580046534 podStartE2EDuration="9.580046534s" podCreationTimestamp="2025-12-01 08:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:41.481789087 +0000 UTC m=+1133.470847008" watchObservedRunningTime="2025-12-01 08:36:41.580046534 +0000 UTC m=+1133.569104455" Dec 01 08:36:41 crc kubenswrapper[4744]: I1201 08:36:41.829737 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:41 crc kubenswrapper[4744]: W1201 08:36:41.832092 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45050404_4ef1_4acb_9bdf_ca9a821fdae7.slice/crio-8d8ec0ef0181aa2badc8540c8c59588591efd1f0d6bf595c15d8f9a9f60e83e2 WatchSource:0}: Error finding container 8d8ec0ef0181aa2badc8540c8c59588591efd1f0d6bf595c15d8f9a9f60e83e2: Status 404 returned error can't find the container with id 8d8ec0ef0181aa2badc8540c8c59588591efd1f0d6bf595c15d8f9a9f60e83e2 Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.312113 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19db2e11-d46c-4fec-a7a4-b3afe22af5fa" path="/var/lib/kubelet/pods/19db2e11-d46c-4fec-a7a4-b3afe22af5fa/volumes" Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.313293 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" path="/var/lib/kubelet/pods/33b049d4-4934-46bd-8dff-cf9995702c8d/volumes" Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.313806 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b626810-3162-42db-83b2-9594a48dd935" path="/var/lib/kubelet/pods/3b626810-3162-42db-83b2-9594a48dd935/volumes" Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.316611 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.435828 4744 generic.go:334] "Generic (PLEG): container finished" podID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerID="16a6e0a17275df1778cff2fd6fc4d9e78c3d3c353e53d816498048caf593d379" exitCode=1 Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.436255 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" event={"ID":"9223dbb6-40f1-4946-b5e3-e84a1e25aa26","Type":"ContainerDied","Data":"16a6e0a17275df1778cff2fd6fc4d9e78c3d3c353e53d816498048caf593d379"} Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.436698 4744 scope.go:117] "RemoveContainer" containerID="16a6e0a17275df1778cff2fd6fc4d9e78c3d3c353e53d816498048caf593d379" Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.444993 4744 generic.go:334] "Generic (PLEG): container finished" podID="59632a6c-e03f-45dd-a08a-561603cb97b5" containerID="d8419e9a848a43bb0186359d12ee3dcf0e7cd46a464cf05568b1c2eb0ca7556e" exitCode=0 Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.445086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" event={"ID":"59632a6c-e03f-45dd-a08a-561603cb97b5","Type":"ContainerDied","Data":"d8419e9a848a43bb0186359d12ee3dcf0e7cd46a464cf05568b1c2eb0ca7556e"} Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.458225 4744 generic.go:334] "Generic (PLEG): container finished" podID="606662e8-bd00-4a68-ba9f-26b9d93d175b" containerID="78e0eae410b90b6f7bcfb7f22d1b6e9e69f6b122e21b37edf7d6df78fe8e20d1" exitCode=0 Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.458317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5fv45" event={"ID":"606662e8-bd00-4a68-ba9f-26b9d93d175b","Type":"ContainerDied","Data":"78e0eae410b90b6f7bcfb7f22d1b6e9e69f6b122e21b37edf7d6df78fe8e20d1"} Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.469061 4744 generic.go:334] "Generic (PLEG): container finished" podID="7230ec92-1f41-491b-a2b5-23aa29cf4d27" containerID="c5449e1ced9108117cf357cc5cde6963d71935c25dddf4f9c40fdfbab1f0aedb" exitCode=0 Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.469132 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f431-account-create-update-lvs6w" event={"ID":"7230ec92-1f41-491b-a2b5-23aa29cf4d27","Type":"ContainerDied","Data":"c5449e1ced9108117cf357cc5cde6963d71935c25dddf4f9c40fdfbab1f0aedb"} Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.487729 4744 generic.go:334] "Generic (PLEG): container finished" podID="65092f5c-34d0-4eaa-9f53-dea23cc222d4" containerID="7f2960ab28ffbc55a4e737a7b3e8ba326b0ddfa06f7cac5faa617e8eb314a7e8" exitCode=0 Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.487798 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" event={"ID":"65092f5c-34d0-4eaa-9f53-dea23cc222d4","Type":"ContainerDied","Data":"7f2960ab28ffbc55a4e737a7b3e8ba326b0ddfa06f7cac5faa617e8eb314a7e8"} Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.493539 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerStarted","Data":"8d8ec0ef0181aa2badc8540c8c59588591efd1f0d6bf595c15d8f9a9f60e83e2"} Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.511902 4744 generic.go:334] "Generic (PLEG): container finished" podID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerID="2af749052de498012cd9a145bbaccbdd326ca98145b2315ea17dcbac1f351917" exitCode=1 Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.512401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56757fd9c7-xswjh" event={"ID":"0dc1a384-3224-46e5-aadc-0421bb72626a","Type":"ContainerDied","Data":"2af749052de498012cd9a145bbaccbdd326ca98145b2315ea17dcbac1f351917"} Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.512590 4744 scope.go:117] "RemoveContainer" containerID="2af749052de498012cd9a145bbaccbdd326ca98145b2315ea17dcbac1f351917" Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.929687 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:42 crc kubenswrapper[4744]: I1201 08:36:42.983259 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.016215 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-operator-scripts\") pod \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.016306 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldq6d\" (UniqueName: \"kubernetes.io/projected/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-kube-api-access-ldq6d\") pod \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\" (UID: \"80d42461-4c7a-4e38-8ac7-1261d5fcb37a\") " Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.017706 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80d42461-4c7a-4e38-8ac7-1261d5fcb37a" (UID: "80d42461-4c7a-4e38-8ac7-1261d5fcb37a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.022056 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-kube-api-access-ldq6d" (OuterVolumeSpecName: "kube-api-access-ldq6d") pod "80d42461-4c7a-4e38-8ac7-1261d5fcb37a" (UID: "80d42461-4c7a-4e38-8ac7-1261d5fcb37a"). InnerVolumeSpecName "kube-api-access-ldq6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.076649 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.102756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.117964 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75s66\" (UniqueName: \"kubernetes.io/projected/e5172715-95f9-48a5-bf83-8fbe12093298-kube-api-access-75s66\") pod \"e5172715-95f9-48a5-bf83-8fbe12093298\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.118484 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5172715-95f9-48a5-bf83-8fbe12093298-operator-scripts\") pod \"e5172715-95f9-48a5-bf83-8fbe12093298\" (UID: \"e5172715-95f9-48a5-bf83-8fbe12093298\") " Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.118946 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5172715-95f9-48a5-bf83-8fbe12093298-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5172715-95f9-48a5-bf83-8fbe12093298" (UID: "e5172715-95f9-48a5-bf83-8fbe12093298"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.119101 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.119125 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldq6d\" (UniqueName: \"kubernetes.io/projected/80d42461-4c7a-4e38-8ac7-1261d5fcb37a-kube-api-access-ldq6d\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.119144 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5172715-95f9-48a5-bf83-8fbe12093298-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.121461 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5172715-95f9-48a5-bf83-8fbe12093298-kube-api-access-75s66" (OuterVolumeSpecName: "kube-api-access-75s66") pod "e5172715-95f9-48a5-bf83-8fbe12093298" (UID: "e5172715-95f9-48a5-bf83-8fbe12093298"). InnerVolumeSpecName "kube-api-access-75s66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.221192 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75s66\" (UniqueName: \"kubernetes.io/projected/e5172715-95f9-48a5-bf83-8fbe12093298-kube-api-access-75s66\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.526285 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sslnq" event={"ID":"e5172715-95f9-48a5-bf83-8fbe12093298","Type":"ContainerDied","Data":"3b95f9a8808eb75db90c4e288399856c08c3060f15d3c6f42fb8623931b65182"} Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.526327 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b95f9a8808eb75db90c4e288399856c08c3060f15d3c6f42fb8623931b65182" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.526390 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sslnq" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.534046 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerStarted","Data":"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5"} Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.547743 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-k46jn" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.547745 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-k46jn" event={"ID":"80d42461-4c7a-4e38-8ac7-1261d5fcb37a","Type":"ContainerDied","Data":"fc72d10a9e48af520b0ce8ddefa472e2588f0301742a0c85026618752f9f2b21"} Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.547867 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc72d10a9e48af520b0ce8ddefa472e2588f0301742a0c85026618752f9f2b21" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.550023 4744 generic.go:334] "Generic (PLEG): container finished" podID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerID="5a759b480179c4427a436f791101d113ac4fa7ef60986e1f4d8b444f307b0a0d" exitCode=1 Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.550083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56757fd9c7-xswjh" event={"ID":"0dc1a384-3224-46e5-aadc-0421bb72626a","Type":"ContainerDied","Data":"5a759b480179c4427a436f791101d113ac4fa7ef60986e1f4d8b444f307b0a0d"} Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.550114 4744 scope.go:117] "RemoveContainer" containerID="2af749052de498012cd9a145bbaccbdd326ca98145b2315ea17dcbac1f351917" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.551064 4744 scope.go:117] "RemoveContainer" containerID="5a759b480179c4427a436f791101d113ac4fa7ef60986e1f4d8b444f307b0a0d" Dec 01 08:36:43 crc kubenswrapper[4744]: E1201 08:36:43.551582 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-56757fd9c7-xswjh_openstack(0dc1a384-3224-46e5-aadc-0421bb72626a)\"" pod="openstack/heat-api-56757fd9c7-xswjh" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.564124 4744 generic.go:334] "Generic (PLEG): container finished" podID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerID="0fed063ad4090c5000900484a23914f5acd8aee381254d4334bdf3b2005412b0" exitCode=1 Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.565399 4744 scope.go:117] "RemoveContainer" containerID="0fed063ad4090c5000900484a23914f5acd8aee381254d4334bdf3b2005412b0" Dec 01 08:36:43 crc kubenswrapper[4744]: E1201 08:36:43.565821 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-f844f6c75-jp9pb_openstack(9223dbb6-40f1-4946-b5e3-e84a1e25aa26)\"" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.566208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" event={"ID":"9223dbb6-40f1-4946-b5e3-e84a1e25aa26","Type":"ContainerDied","Data":"0fed063ad4090c5000900484a23914f5acd8aee381254d4334bdf3b2005412b0"} Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.642533 4744 scope.go:117] "RemoveContainer" containerID="16a6e0a17275df1778cff2fd6fc4d9e78c3d3c353e53d816498048caf593d379" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.857009 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.151:9292/healthcheck\": read tcp 10.217.0.2:45934->10.217.0.151:9292: read: connection reset by peer" Dec 01 08:36:43 crc kubenswrapper[4744]: I1201 08:36:43.858013 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9292/healthcheck\": read tcp 10.217.0.2:45926->10.217.0.151:9292: read: connection reset by peer" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.042746 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.107606 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.145153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7230ec92-1f41-491b-a2b5-23aa29cf4d27-operator-scripts\") pod \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.145335 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5zwg\" (UniqueName: \"kubernetes.io/projected/7230ec92-1f41-491b-a2b5-23aa29cf4d27-kube-api-access-n5zwg\") pod \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\" (UID: \"7230ec92-1f41-491b-a2b5-23aa29cf4d27\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.148216 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7230ec92-1f41-491b-a2b5-23aa29cf4d27-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7230ec92-1f41-491b-a2b5-23aa29cf4d27" (UID: "7230ec92-1f41-491b-a2b5-23aa29cf4d27"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.155616 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7230ec92-1f41-491b-a2b5-23aa29cf4d27-kube-api-access-n5zwg" (OuterVolumeSpecName: "kube-api-access-n5zwg") pod "7230ec92-1f41-491b-a2b5-23aa29cf4d27" (UID: "7230ec92-1f41-491b-a2b5-23aa29cf4d27"). InnerVolumeSpecName "kube-api-access-n5zwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.249705 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5zwg\" (UniqueName: \"kubernetes.io/projected/7230ec92-1f41-491b-a2b5-23aa29cf4d27-kube-api-access-n5zwg\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.249734 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7230ec92-1f41-491b-a2b5-23aa29cf4d27-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.276480 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.329727 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.344000 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.454817 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjdmw\" (UniqueName: \"kubernetes.io/projected/65092f5c-34d0-4eaa-9f53-dea23cc222d4-kube-api-access-wjdmw\") pod \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.455713 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/606662e8-bd00-4a68-ba9f-26b9d93d175b-operator-scripts\") pod \"606662e8-bd00-4a68-ba9f-26b9d93d175b\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.455833 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcrgj\" (UniqueName: \"kubernetes.io/projected/606662e8-bd00-4a68-ba9f-26b9d93d175b-kube-api-access-hcrgj\") pod \"606662e8-bd00-4a68-ba9f-26b9d93d175b\" (UID: \"606662e8-bd00-4a68-ba9f-26b9d93d175b\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.455926 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7s4m\" (UniqueName: \"kubernetes.io/projected/59632a6c-e03f-45dd-a08a-561603cb97b5-kube-api-access-x7s4m\") pod \"59632a6c-e03f-45dd-a08a-561603cb97b5\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.456087 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59632a6c-e03f-45dd-a08a-561603cb97b5-operator-scripts\") pod \"59632a6c-e03f-45dd-a08a-561603cb97b5\" (UID: \"59632a6c-e03f-45dd-a08a-561603cb97b5\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.456209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65092f5c-34d0-4eaa-9f53-dea23cc222d4-operator-scripts\") pod \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\" (UID: \"65092f5c-34d0-4eaa-9f53-dea23cc222d4\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.456917 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/606662e8-bd00-4a68-ba9f-26b9d93d175b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "606662e8-bd00-4a68-ba9f-26b9d93d175b" (UID: "606662e8-bd00-4a68-ba9f-26b9d93d175b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.457451 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59632a6c-e03f-45dd-a08a-561603cb97b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "59632a6c-e03f-45dd-a08a-561603cb97b5" (UID: "59632a6c-e03f-45dd-a08a-561603cb97b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.457719 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/606662e8-bd00-4a68-ba9f-26b9d93d175b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.458095 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65092f5c-34d0-4eaa-9f53-dea23cc222d4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65092f5c-34d0-4eaa-9f53-dea23cc222d4" (UID: "65092f5c-34d0-4eaa-9f53-dea23cc222d4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.465451 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/606662e8-bd00-4a68-ba9f-26b9d93d175b-kube-api-access-hcrgj" (OuterVolumeSpecName: "kube-api-access-hcrgj") pod "606662e8-bd00-4a68-ba9f-26b9d93d175b" (UID: "606662e8-bd00-4a68-ba9f-26b9d93d175b"). InnerVolumeSpecName "kube-api-access-hcrgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.465730 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65092f5c-34d0-4eaa-9f53-dea23cc222d4-kube-api-access-wjdmw" (OuterVolumeSpecName: "kube-api-access-wjdmw") pod "65092f5c-34d0-4eaa-9f53-dea23cc222d4" (UID: "65092f5c-34d0-4eaa-9f53-dea23cc222d4"). InnerVolumeSpecName "kube-api-access-wjdmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.468641 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59632a6c-e03f-45dd-a08a-561603cb97b5-kube-api-access-x7s4m" (OuterVolumeSpecName: "kube-api-access-x7s4m") pod "59632a6c-e03f-45dd-a08a-561603cb97b5" (UID: "59632a6c-e03f-45dd-a08a-561603cb97b5"). InnerVolumeSpecName "kube-api-access-x7s4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.559154 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59632a6c-e03f-45dd-a08a-561603cb97b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.560051 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65092f5c-34d0-4eaa-9f53-dea23cc222d4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.560131 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjdmw\" (UniqueName: \"kubernetes.io/projected/65092f5c-34d0-4eaa-9f53-dea23cc222d4-kube-api-access-wjdmw\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.560188 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcrgj\" (UniqueName: \"kubernetes.io/projected/606662e8-bd00-4a68-ba9f-26b9d93d175b-kube-api-access-hcrgj\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.560250 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7s4m\" (UniqueName: \"kubernetes.io/projected/59632a6c-e03f-45dd-a08a-561603cb97b5-kube-api-access-x7s4m\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.610612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" event={"ID":"65092f5c-34d0-4eaa-9f53-dea23cc222d4","Type":"ContainerDied","Data":"dafdc58a87e0394a1efcba0b6638f52e16c412bd723648368d793eca89ae7a7b"} Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.611229 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dafdc58a87e0394a1efcba0b6638f52e16c412bd723648368d793eca89ae7a7b" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.611193 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fe16-account-create-update-h8bdl" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.623805 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerStarted","Data":"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f"} Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.635387 4744 scope.go:117] "RemoveContainer" containerID="5a759b480179c4427a436f791101d113ac4fa7ef60986e1f4d8b444f307b0a0d" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.635988 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-56757fd9c7-xswjh_openstack(0dc1a384-3224-46e5-aadc-0421bb72626a)\"" pod="openstack/heat-api-56757fd9c7-xswjh" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.643572 4744 scope.go:117] "RemoveContainer" containerID="0fed063ad4090c5000900484a23914f5acd8aee381254d4334bdf3b2005412b0" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.643785 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-f844f6c75-jp9pb_openstack(9223dbb6-40f1-4946-b5e3-e84a1e25aa26)\"" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.647011 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" event={"ID":"59632a6c-e03f-45dd-a08a-561603cb97b5","Type":"ContainerDied","Data":"8d2a7d55cbedc275d40c187a8ecba14645188463665af7caa30638aae9ce853f"} Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.647197 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d2a7d55cbedc275d40c187a8ecba14645188463665af7caa30638aae9ce853f" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.647337 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8cec-account-create-update-xlrzh" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.654292 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5fv45" event={"ID":"606662e8-bd00-4a68-ba9f-26b9d93d175b","Type":"ContainerDied","Data":"80595464f83be08d3bc42ceabc93e31d0c229295156fa6a45c8bfedd55791212"} Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.654347 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80595464f83be08d3bc42ceabc93e31d0c229295156fa6a45c8bfedd55791212" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.654437 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5fv45" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.661067 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f431-account-create-update-lvs6w" event={"ID":"7230ec92-1f41-491b-a2b5-23aa29cf4d27","Type":"ContainerDied","Data":"77c075234dc60dfd106f57d60f4cb94174a2a77e6091efaae1bb8936c9501919"} Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.661110 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77c075234dc60dfd106f57d60f4cb94174a2a77e6091efaae1bb8936c9501919" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.661198 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f431-account-create-update-lvs6w" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.669127 4744 generic.go:334] "Generic (PLEG): container finished" podID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerID="2411f961caa796d67d32b9f30533c1102ca0802392405051a9b1df63f83097a2" exitCode=0 Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.669170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a0a609f-e643-4e6b-b0f8-428fd792fce9","Type":"ContainerDied","Data":"2411f961caa796d67d32b9f30533c1102ca0802392405051a9b1df63f83097a2"} Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.713291 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.884053 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-internal-tls-certs\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.884335 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-combined-ca-bundle\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.884469 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-httpd-run\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.884615 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-scripts\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.884726 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-logs\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.884897 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzng5\" (UniqueName: \"kubernetes.io/projected/4a0a609f-e643-4e6b-b0f8-428fd792fce9-kube-api-access-fzng5\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.884987 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.885073 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-config-data\") pod \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\" (UID: \"4a0a609f-e643-4e6b-b0f8-428fd792fce9\") " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.888283 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-logs" (OuterVolumeSpecName: "logs") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.888903 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.891953 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a0a609f-e643-4e6b-b0f8-428fd792fce9-kube-api-access-fzng5" (OuterVolumeSpecName: "kube-api-access-fzng5") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "kube-api-access-fzng5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.893577 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-scripts" (OuterVolumeSpecName: "scripts") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.897229 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.934768 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t86st"] Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.935202 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-log" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.935214 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-log" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.935234 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65092f5c-34d0-4eaa-9f53-dea23cc222d4" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.935241 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="65092f5c-34d0-4eaa-9f53-dea23cc222d4" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.935253 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="606662e8-bd00-4a68-ba9f-26b9d93d175b" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.935259 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="606662e8-bd00-4a68-ba9f-26b9d93d175b" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.935268 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7230ec92-1f41-491b-a2b5-23aa29cf4d27" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936230 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936618 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7230ec92-1f41-491b-a2b5-23aa29cf4d27" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.936681 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5172715-95f9-48a5-bf83-8fbe12093298" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936689 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5172715-95f9-48a5-bf83-8fbe12093298" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.936708 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerName="heat-cfnapi" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936714 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerName="heat-cfnapi" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.936725 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59632a6c-e03f-45dd-a08a-561603cb97b5" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936731 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="59632a6c-e03f-45dd-a08a-561603cb97b5" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.936742 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d42461-4c7a-4e38-8ac7-1261d5fcb37a" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936748 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d42461-4c7a-4e38-8ac7-1261d5fcb37a" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: E1201 08:36:44.936758 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-httpd" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936764 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-httpd" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.936991 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-httpd" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937005 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" containerName="glance-log" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937017 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="59632a6c-e03f-45dd-a08a-561603cb97b5" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937028 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="80d42461-4c7a-4e38-8ac7-1261d5fcb37a" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937039 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="65092f5c-34d0-4eaa-9f53-dea23cc222d4" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937051 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="606662e8-bd00-4a68-ba9f-26b9d93d175b" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937061 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b049d4-4934-46bd-8dff-cf9995702c8d" containerName="heat-cfnapi" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937074 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7230ec92-1f41-491b-a2b5-23aa29cf4d27" containerName="mariadb-account-create-update" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937081 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5172715-95f9-48a5-bf83-8fbe12093298" containerName="mariadb-database-create" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.937741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.944018 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s85xn" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.945895 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.948062 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.961670 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t86st"] Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.973725 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-config-data" (OuterVolumeSpecName: "config-data") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.987636 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.987672 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.987680 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzng5\" (UniqueName: \"kubernetes.io/projected/4a0a609f-e643-4e6b-b0f8-428fd792fce9-kube-api-access-fzng5\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.987700 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.987709 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.987718 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:44 crc kubenswrapper[4744]: I1201 08:36:44.987727 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a0a609f-e643-4e6b-b0f8-428fd792fce9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.034511 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4a0a609f-e643-4e6b-b0f8-428fd792fce9" (UID: "4a0a609f-e643-4e6b-b0f8-428fd792fce9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.054645 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.111074 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-config-data\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.119585 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.123310 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w2wl\" (UniqueName: \"kubernetes.io/projected/522db2e5-3aed-45b0-a228-ab2b875a3cee-kube-api-access-5w2wl\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.127371 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-scripts\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.131438 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.131796 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a0a609f-e643-4e6b-b0f8-428fd792fce9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.232765 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w2wl\" (UniqueName: \"kubernetes.io/projected/522db2e5-3aed-45b0-a228-ab2b875a3cee-kube-api-access-5w2wl\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.232843 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-scripts\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.232886 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-config-data\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.232966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.236172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-scripts\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.247563 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.247631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-config-data\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.249927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w2wl\" (UniqueName: \"kubernetes.io/projected/522db2e5-3aed-45b0-a228-ab2b875a3cee-kube-api-access-5w2wl\") pod \"nova-cell0-conductor-db-sync-t86st\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.277214 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.679882 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerStarted","Data":"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e"} Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.682944 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4a0a609f-e643-4e6b-b0f8-428fd792fce9","Type":"ContainerDied","Data":"6c3f52aa05b946ca3ba4e7c1e13f59c75f94872788bff36a88ff53951d745b98"} Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.682971 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.683001 4744 scope.go:117] "RemoveContainer" containerID="2411f961caa796d67d32b9f30533c1102ca0802392405051a9b1df63f83097a2" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.683552 4744 scope.go:117] "RemoveContainer" containerID="0fed063ad4090c5000900484a23914f5acd8aee381254d4334bdf3b2005412b0" Dec 01 08:36:45 crc kubenswrapper[4744]: E1201 08:36:45.683920 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-f844f6c75-jp9pb_openstack(9223dbb6-40f1-4946-b5e3-e84a1e25aa26)\"" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.715248 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.723352 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.739501 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.739715 4744 scope.go:117] "RemoveContainer" containerID="3e9efe156cc4be4376971c5a296054f22c889d63d7a2cf1180f1f935b8ec5791" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.741520 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.745809 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.746047 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gl4sq" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.746565 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.746705 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.752719 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.802905 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t86st"] Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.844809 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5c1409-3b79-4680-beb3-14ffcf0fb055-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.844892 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.844956 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.844992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.845099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.845126 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5c1409-3b79-4680-beb3-14ffcf0fb055-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.845157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-988ws\" (UniqueName: \"kubernetes.io/projected/7b5c1409-3b79-4680-beb3-14ffcf0fb055-kube-api-access-988ws\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.845203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947434 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947575 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947615 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5c1409-3b79-4680-beb3-14ffcf0fb055-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947653 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-988ws\" (UniqueName: \"kubernetes.io/projected/7b5c1409-3b79-4680-beb3-14ffcf0fb055-kube-api-access-988ws\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947691 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947713 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.947985 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5c1409-3b79-4680-beb3-14ffcf0fb055-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.948273 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5c1409-3b79-4680-beb3-14ffcf0fb055-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.948440 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5c1409-3b79-4680-beb3-14ffcf0fb055-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.953248 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.956129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.957813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.963881 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5c1409-3b79-4680-beb3-14ffcf0fb055-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.964538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-988ws\" (UniqueName: \"kubernetes.io/projected/7b5c1409-3b79-4680-beb3-14ffcf0fb055-kube-api-access-988ws\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:45 crc kubenswrapper[4744]: I1201 08:36:45.985028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"7b5c1409-3b79-4680-beb3-14ffcf0fb055\") " pod="openstack/glance-default-internal-api-0" Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.078583 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.314390 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a0a609f-e643-4e6b-b0f8-428fd792fce9" path="/var/lib/kubelet/pods/4a0a609f-e643-4e6b-b0f8-428fd792fce9/volumes" Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.657108 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.694719 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b5c1409-3b79-4680-beb3-14ffcf0fb055","Type":"ContainerStarted","Data":"82a640d31853c4168199fb1f8a4eea75f5f1aebac59d1f621cc63216739292c5"} Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.704108 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerStarted","Data":"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903"} Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.704592 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-central-agent" containerID="cri-o://11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" gracePeriod=30 Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.704713 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.705252 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="proxy-httpd" containerID="cri-o://5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" gracePeriod=30 Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.705340 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="sg-core" containerID="cri-o://fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" gracePeriod=30 Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.705441 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-notification-agent" containerID="cri-o://68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" gracePeriod=30 Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.710455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t86st" event={"ID":"522db2e5-3aed-45b0-a228-ab2b875a3cee","Type":"ContainerStarted","Data":"9efe2ac4dbde65e1061d55e8d1e051b23e3ffd47403fdbf4101cff3264afe796"} Dec 01 08:36:46 crc kubenswrapper[4744]: I1201 08:36:46.731953 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.435608173 podStartE2EDuration="6.731935089s" podCreationTimestamp="2025-12-01 08:36:40 +0000 UTC" firstStartedPulling="2025-12-01 08:36:41.83524288 +0000 UTC m=+1133.824300801" lastFinishedPulling="2025-12-01 08:36:46.131569796 +0000 UTC m=+1138.120627717" observedRunningTime="2025-12-01 08:36:46.724813197 +0000 UTC m=+1138.713871118" watchObservedRunningTime="2025-12-01 08:36:46.731935089 +0000 UTC m=+1138.720993010" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.544587 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.688201 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-scripts\") pod \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.688630 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-sg-core-conf-yaml\") pod \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.688676 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-run-httpd\") pod \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.688737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-config-data\") pod \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.688779 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-log-httpd\") pod \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.688814 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-combined-ca-bundle\") pod \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.689168 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "45050404-4ef1-4acb-9bdf-ca9a821fdae7" (UID: "45050404-4ef1-4acb-9bdf-ca9a821fdae7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.689209 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "45050404-4ef1-4acb-9bdf-ca9a821fdae7" (UID: "45050404-4ef1-4acb-9bdf-ca9a821fdae7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.689234 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zfkm\" (UniqueName: \"kubernetes.io/projected/45050404-4ef1-4acb-9bdf-ca9a821fdae7-kube-api-access-5zfkm\") pod \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\" (UID: \"45050404-4ef1-4acb-9bdf-ca9a821fdae7\") " Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.689807 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.689829 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45050404-4ef1-4acb-9bdf-ca9a821fdae7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.692518 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45050404-4ef1-4acb-9bdf-ca9a821fdae7-kube-api-access-5zfkm" (OuterVolumeSpecName: "kube-api-access-5zfkm") pod "45050404-4ef1-4acb-9bdf-ca9a821fdae7" (UID: "45050404-4ef1-4acb-9bdf-ca9a821fdae7"). InnerVolumeSpecName "kube-api-access-5zfkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.694284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-scripts" (OuterVolumeSpecName: "scripts") pod "45050404-4ef1-4acb-9bdf-ca9a821fdae7" (UID: "45050404-4ef1-4acb-9bdf-ca9a821fdae7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.728919 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "45050404-4ef1-4acb-9bdf-ca9a821fdae7" (UID: "45050404-4ef1-4acb-9bdf-ca9a821fdae7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.747591 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b5c1409-3b79-4680-beb3-14ffcf0fb055","Type":"ContainerStarted","Data":"3cc69aed86e0fd28fd460e33763c337fea0055ce1577dd05178e90d59ef2043c"} Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756388 4744 generic.go:334] "Generic (PLEG): container finished" podID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerID="5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" exitCode=0 Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756440 4744 generic.go:334] "Generic (PLEG): container finished" podID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerID="fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" exitCode=2 Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756449 4744 generic.go:334] "Generic (PLEG): container finished" podID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerID="68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" exitCode=0 Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756455 4744 generic.go:334] "Generic (PLEG): container finished" podID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerID="11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" exitCode=0 Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756474 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerDied","Data":"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903"} Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756497 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerDied","Data":"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e"} Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerDied","Data":"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f"} Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756521 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerDied","Data":"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5"} Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756531 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45050404-4ef1-4acb-9bdf-ca9a821fdae7","Type":"ContainerDied","Data":"8d8ec0ef0181aa2badc8540c8c59588591efd1f0d6bf595c15d8f9a9f60e83e2"} Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756545 4744 scope.go:117] "RemoveContainer" containerID="5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.756671 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.791671 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zfkm\" (UniqueName: \"kubernetes.io/projected/45050404-4ef1-4acb-9bdf-ca9a821fdae7-kube-api-access-5zfkm\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.791694 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.791703 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.798126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45050404-4ef1-4acb-9bdf-ca9a821fdae7" (UID: "45050404-4ef1-4acb-9bdf-ca9a821fdae7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.812253 4744 scope.go:117] "RemoveContainer" containerID="fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.815470 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-config-data" (OuterVolumeSpecName: "config-data") pod "45050404-4ef1-4acb-9bdf-ca9a821fdae7" (UID: "45050404-4ef1-4acb-9bdf-ca9a821fdae7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.832381 4744 scope.go:117] "RemoveContainer" containerID="68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.860221 4744 scope.go:117] "RemoveContainer" containerID="11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.880455 4744 scope.go:117] "RemoveContainer" containerID="5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" Dec 01 08:36:47 crc kubenswrapper[4744]: E1201 08:36:47.880996 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": container with ID starting with 5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903 not found: ID does not exist" containerID="5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.881077 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903"} err="failed to get container status \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": rpc error: code = NotFound desc = could not find container \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": container with ID starting with 5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.881104 4744 scope.go:117] "RemoveContainer" containerID="fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" Dec 01 08:36:47 crc kubenswrapper[4744]: E1201 08:36:47.881763 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": container with ID starting with fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e not found: ID does not exist" containerID="fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.881781 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e"} err="failed to get container status \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": rpc error: code = NotFound desc = could not find container \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": container with ID starting with fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.881793 4744 scope.go:117] "RemoveContainer" containerID="68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" Dec 01 08:36:47 crc kubenswrapper[4744]: E1201 08:36:47.882056 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": container with ID starting with 68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f not found: ID does not exist" containerID="68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.882077 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f"} err="failed to get container status \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": rpc error: code = NotFound desc = could not find container \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": container with ID starting with 68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.882088 4744 scope.go:117] "RemoveContainer" containerID="11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" Dec 01 08:36:47 crc kubenswrapper[4744]: E1201 08:36:47.882344 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": container with ID starting with 11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5 not found: ID does not exist" containerID="11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.882361 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5"} err="failed to get container status \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": rpc error: code = NotFound desc = could not find container \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": container with ID starting with 11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.882374 4744 scope.go:117] "RemoveContainer" containerID="5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.882850 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903"} err="failed to get container status \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": rpc error: code = NotFound desc = could not find container \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": container with ID starting with 5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.882889 4744 scope.go:117] "RemoveContainer" containerID="fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.883243 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e"} err="failed to get container status \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": rpc error: code = NotFound desc = could not find container \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": container with ID starting with fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.883264 4744 scope.go:117] "RemoveContainer" containerID="68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.883734 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f"} err="failed to get container status \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": rpc error: code = NotFound desc = could not find container \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": container with ID starting with 68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.883791 4744 scope.go:117] "RemoveContainer" containerID="11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884030 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5"} err="failed to get container status \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": rpc error: code = NotFound desc = could not find container \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": container with ID starting with 11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884046 4744 scope.go:117] "RemoveContainer" containerID="5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884329 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903"} err="failed to get container status \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": rpc error: code = NotFound desc = could not find container \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": container with ID starting with 5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884345 4744 scope.go:117] "RemoveContainer" containerID="fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884666 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e"} err="failed to get container status \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": rpc error: code = NotFound desc = could not find container \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": container with ID starting with fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884681 4744 scope.go:117] "RemoveContainer" containerID="68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884895 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f"} err="failed to get container status \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": rpc error: code = NotFound desc = could not find container \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": container with ID starting with 68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.884910 4744 scope.go:117] "RemoveContainer" containerID="11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.885124 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5"} err="failed to get container status \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": rpc error: code = NotFound desc = could not find container \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": container with ID starting with 11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.885140 4744 scope.go:117] "RemoveContainer" containerID="5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.885459 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903"} err="failed to get container status \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": rpc error: code = NotFound desc = could not find container \"5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903\": container with ID starting with 5d15eb7529ea7242e4d3b3b3f3bc3131bdbca4b012fb82c19280d9327d242903 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.885509 4744 scope.go:117] "RemoveContainer" containerID="fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.885880 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e"} err="failed to get container status \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": rpc error: code = NotFound desc = could not find container \"fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e\": container with ID starting with fe4436d1cd98b0f0ffe2da89480c8b7f61f25bf149a6b62627084ff6e516f03e not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.885898 4744 scope.go:117] "RemoveContainer" containerID="68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.887042 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f"} err="failed to get container status \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": rpc error: code = NotFound desc = could not find container \"68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f\": container with ID starting with 68ee93154bffa70d22234dd9c4e4189fb3810d60345045a1f22785226ed0e65f not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.887071 4744 scope.go:117] "RemoveContainer" containerID="11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.887399 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5"} err="failed to get container status \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": rpc error: code = NotFound desc = could not find container \"11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5\": container with ID starting with 11f1c5d93ef26e0a34c4a0a747fb54f6caddd0e8d4564fd8072e3f532fc20eb5 not found: ID does not exist" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.893963 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:47 crc kubenswrapper[4744]: I1201 08:36:47.893988 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45050404-4ef1-4acb-9bdf-ca9a821fdae7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.076555 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.077641 4744 scope.go:117] "RemoveContainer" containerID="5a759b480179c4427a436f791101d113ac4fa7ef60986e1f4d8b444f307b0a0d" Dec 01 08:36:48 crc kubenswrapper[4744]: E1201 08:36:48.077957 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-56757fd9c7-xswjh_openstack(0dc1a384-3224-46e5-aadc-0421bb72626a)\"" pod="openstack/heat-api-56757fd9c7-xswjh" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.087008 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.092625 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.102490 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.102528 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.103174 4744 scope.go:117] "RemoveContainer" containerID="0fed063ad4090c5000900484a23914f5acd8aee381254d4334bdf3b2005412b0" Dec 01 08:36:48 crc kubenswrapper[4744]: E1201 08:36:48.103368 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-f844f6c75-jp9pb_openstack(9223dbb6-40f1-4946-b5e3-e84a1e25aa26)\"" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.105849 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.118590 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:48 crc kubenswrapper[4744]: E1201 08:36:48.119058 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-central-agent" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119077 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-central-agent" Dec 01 08:36:48 crc kubenswrapper[4744]: E1201 08:36:48.119090 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-notification-agent" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119096 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-notification-agent" Dec 01 08:36:48 crc kubenswrapper[4744]: E1201 08:36:48.119114 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="sg-core" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119119 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="sg-core" Dec 01 08:36:48 crc kubenswrapper[4744]: E1201 08:36:48.119140 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="proxy-httpd" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119148 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="proxy-httpd" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119307 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="proxy-httpd" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119329 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="sg-core" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119340 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-notification-agent" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.119353 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" containerName="ceilometer-central-agent" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.120932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.124918 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.125272 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.131014 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.297833 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45050404-4ef1-4acb-9bdf-ca9a821fdae7" path="/var/lib/kubelet/pods/45050404-4ef1-4acb-9bdf-ca9a821fdae7/volumes" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.303594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-run-httpd\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.303626 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-scripts\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.303650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zkj9\" (UniqueName: \"kubernetes.io/projected/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-kube-api-access-5zkj9\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.303678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.303829 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-log-httpd\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.303925 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.304066 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-config-data\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.407958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.408019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-log-httpd\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.408075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.408185 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-config-data\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.408291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-run-httpd\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.408305 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-scripts\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.408323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zkj9\" (UniqueName: \"kubernetes.io/projected/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-kube-api-access-5zkj9\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.409060 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-run-httpd\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.410448 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-log-httpd\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.412665 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.412988 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.423160 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.425198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.426212 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-scripts\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.426439 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-config-data\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.426763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zkj9\" (UniqueName: \"kubernetes.io/projected/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-kube-api-access-5zkj9\") pod \"ceilometer-0\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.436569 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.767715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b5c1409-3b79-4680-beb3-14ffcf0fb055","Type":"ContainerStarted","Data":"f2828a03c17feec984b2504f297f2e80d6ff9636ce9b4119599548c2f07a17dc"} Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.768265 4744 scope.go:117] "RemoveContainer" containerID="5a759b480179c4427a436f791101d113ac4fa7ef60986e1f4d8b444f307b0a0d" Dec 01 08:36:48 crc kubenswrapper[4744]: E1201 08:36:48.768589 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-56757fd9c7-xswjh_openstack(0dc1a384-3224-46e5-aadc-0421bb72626a)\"" pod="openstack/heat-api-56757fd9c7-xswjh" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.791452 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.791435159 podStartE2EDuration="3.791435159s" podCreationTimestamp="2025-12-01 08:36:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:36:48.783847293 +0000 UTC m=+1140.772905214" watchObservedRunningTime="2025-12-01 08:36:48.791435159 +0000 UTC m=+1140.780493080" Dec 01 08:36:48 crc kubenswrapper[4744]: I1201 08:36:48.897171 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:49 crc kubenswrapper[4744]: I1201 08:36:49.777099 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerStarted","Data":"d1812ea927ee5251a8258bea5e90039cdd8d48c315e01fec0bd5243979d76070"} Dec 01 08:36:49 crc kubenswrapper[4744]: I1201 08:36:49.917517 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:36:50 crc kubenswrapper[4744]: I1201 08:36:50.788642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerStarted","Data":"ac755322f88c3b6773a34838f8e0ac6eb4fb0e19d53c5de6f433425af4916fa9"} Dec 01 08:36:50 crc kubenswrapper[4744]: I1201 08:36:50.788928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerStarted","Data":"15da7c51c0b82a6e91e1bc9596def580cddaa82df9899537a23d4933f357da05"} Dec 01 08:36:51 crc kubenswrapper[4744]: I1201 08:36:51.114308 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-68d7bb6659-xxvcz" Dec 01 08:36:51 crc kubenswrapper[4744]: I1201 08:36:51.131643 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5c9d7d48-5qkmn" Dec 01 08:36:51 crc kubenswrapper[4744]: I1201 08:36:51.194526 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f844f6c75-jp9pb"] Dec 01 08:36:51 crc kubenswrapper[4744]: I1201 08:36:51.202252 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-56757fd9c7-xswjh"] Dec 01 08:36:53 crc kubenswrapper[4744]: I1201 08:36:53.072984 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5bf9f8b584-7k2r7" Dec 01 08:36:53 crc kubenswrapper[4744]: I1201 08:36:53.122837 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5b7d595658-kpqqh"] Dec 01 08:36:53 crc kubenswrapper[4744]: I1201 08:36:53.123030 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-5b7d595658-kpqqh" podUID="947af6f3-b9a2-46da-98d3-bed308f4ba29" containerName="heat-engine" containerID="cri-o://a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" gracePeriod=60 Dec 01 08:36:54 crc kubenswrapper[4744]: E1201 08:36:54.054482 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 08:36:54 crc kubenswrapper[4744]: E1201 08:36:54.057640 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 08:36:54 crc kubenswrapper[4744]: E1201 08:36:54.059878 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 08:36:54 crc kubenswrapper[4744]: E1201 08:36:54.059917 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5b7d595658-kpqqh" podUID="947af6f3-b9a2-46da-98d3-bed308f4ba29" containerName="heat-engine" Dec 01 08:36:56 crc kubenswrapper[4744]: I1201 08:36:56.079228 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:56 crc kubenswrapper[4744]: I1201 08:36:56.079521 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:56 crc kubenswrapper[4744]: I1201 08:36:56.120082 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:56 crc kubenswrapper[4744]: I1201 08:36:56.139241 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:56 crc kubenswrapper[4744]: I1201 08:36:56.874517 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:56 crc kubenswrapper[4744]: I1201 08:36:56.874963 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.684171 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.723134 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.788861 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-combined-ca-bundle\") pod \"0dc1a384-3224-46e5-aadc-0421bb72626a\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.788960 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx2m5\" (UniqueName: \"kubernetes.io/projected/0dc1a384-3224-46e5-aadc-0421bb72626a-kube-api-access-kx2m5\") pod \"0dc1a384-3224-46e5-aadc-0421bb72626a\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.788988 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data\") pod \"0dc1a384-3224-46e5-aadc-0421bb72626a\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.789039 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data-custom\") pod \"0dc1a384-3224-46e5-aadc-0421bb72626a\" (UID: \"0dc1a384-3224-46e5-aadc-0421bb72626a\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.800139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0dc1a384-3224-46e5-aadc-0421bb72626a" (UID: "0dc1a384-3224-46e5-aadc-0421bb72626a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.809881 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dc1a384-3224-46e5-aadc-0421bb72626a-kube-api-access-kx2m5" (OuterVolumeSpecName: "kube-api-access-kx2m5") pod "0dc1a384-3224-46e5-aadc-0421bb72626a" (UID: "0dc1a384-3224-46e5-aadc-0421bb72626a"). InnerVolumeSpecName "kube-api-access-kx2m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.889885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dc1a384-3224-46e5-aadc-0421bb72626a" (UID: "0dc1a384-3224-46e5-aadc-0421bb72626a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.891511 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data-custom\") pod \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.891627 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ht8pb\" (UniqueName: \"kubernetes.io/projected/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-kube-api-access-ht8pb\") pod \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.891657 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data\") pod \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.891883 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-combined-ca-bundle\") pod \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\" (UID: \"9223dbb6-40f1-4946-b5e3-e84a1e25aa26\") " Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.892869 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.892894 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx2m5\" (UniqueName: \"kubernetes.io/projected/0dc1a384-3224-46e5-aadc-0421bb72626a-kube-api-access-kx2m5\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.892907 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.894268 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data" (OuterVolumeSpecName: "config-data") pod "0dc1a384-3224-46e5-aadc-0421bb72626a" (UID: "0dc1a384-3224-46e5-aadc-0421bb72626a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.905092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerStarted","Data":"9111e09f332b16dd57d30bf177e2e2096e7bb9b9a5c1cd20e2459af23f76740f"} Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.930009 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56757fd9c7-xswjh" event={"ID":"0dc1a384-3224-46e5-aadc-0421bb72626a","Type":"ContainerDied","Data":"6ab0f5f23dac41b15548145cf5129567217b3d10d5f301f43e982d66df7db1c5"} Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.930292 4744 scope.go:117] "RemoveContainer" containerID="5a759b480179c4427a436f791101d113ac4fa7ef60986e1f4d8b444f307b0a0d" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.930625 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56757fd9c7-xswjh" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.932945 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-kube-api-access-ht8pb" (OuterVolumeSpecName: "kube-api-access-ht8pb") pod "9223dbb6-40f1-4946-b5e3-e84a1e25aa26" (UID: "9223dbb6-40f1-4946-b5e3-e84a1e25aa26"). InnerVolumeSpecName "kube-api-access-ht8pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.933362 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9223dbb6-40f1-4946-b5e3-e84a1e25aa26" (UID: "9223dbb6-40f1-4946-b5e3-e84a1e25aa26"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.936340 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t86st" event={"ID":"522db2e5-3aed-45b0-a228-ab2b875a3cee","Type":"ContainerStarted","Data":"f12d78059710426c963a1cb1d2f3860cc3f000b0f18373f3a8687e7477a5cd6d"} Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.940846 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" event={"ID":"9223dbb6-40f1-4946-b5e3-e84a1e25aa26","Type":"ContainerDied","Data":"ead31e7842ab61d4413f0f673306fb4b253db0a8b28e5c6e19b70a3a02683df8"} Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.942548 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f844f6c75-jp9pb" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.959007 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-t86st" podStartSLOduration=2.285432027 podStartE2EDuration="13.958983319s" podCreationTimestamp="2025-12-01 08:36:44 +0000 UTC" firstStartedPulling="2025-12-01 08:36:45.824632776 +0000 UTC m=+1137.813690697" lastFinishedPulling="2025-12-01 08:36:57.498184068 +0000 UTC m=+1149.487241989" observedRunningTime="2025-12-01 08:36:57.95272072 +0000 UTC m=+1149.941778641" watchObservedRunningTime="2025-12-01 08:36:57.958983319 +0000 UTC m=+1149.948041240" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.960235 4744 scope.go:117] "RemoveContainer" containerID="0fed063ad4090c5000900484a23914f5acd8aee381254d4334bdf3b2005412b0" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.978330 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9223dbb6-40f1-4946-b5e3-e84a1e25aa26" (UID: "9223dbb6-40f1-4946-b5e3-e84a1e25aa26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.989314 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-56757fd9c7-xswjh"] Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.994107 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc1a384-3224-46e5-aadc-0421bb72626a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.994141 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.994152 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:57 crc kubenswrapper[4744]: I1201 08:36:57.994161 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ht8pb\" (UniqueName: \"kubernetes.io/projected/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-kube-api-access-ht8pb\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:57.999981 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-56757fd9c7-xswjh"] Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.007051 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data" (OuterVolumeSpecName: "config-data") pod "9223dbb6-40f1-4946-b5e3-e84a1e25aa26" (UID: "9223dbb6-40f1-4946-b5e3-e84a1e25aa26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.096013 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9223dbb6-40f1-4946-b5e3-e84a1e25aa26-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.269623 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f844f6c75-jp9pb"] Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.278021 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-f844f6c75-jp9pb"] Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.295742 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" path="/var/lib/kubelet/pods/0dc1a384-3224-46e5-aadc-0421bb72626a/volumes" Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.296275 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" path="/var/lib/kubelet/pods/9223dbb6-40f1-4946-b5e3-e84a1e25aa26/volumes" Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.942038 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 08:36:58 crc kubenswrapper[4744]: I1201 08:36:58.944673 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 08:37:00 crc kubenswrapper[4744]: I1201 08:37:00.998093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerStarted","Data":"d9e4f2a9231bd4415c72c8bc42a899e4d0cba0bf2dfd0f44aaeb71f401c9303b"} Dec 01 08:37:00 crc kubenswrapper[4744]: I1201 08:37:00.998778 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-central-agent" containerID="cri-o://ac755322f88c3b6773a34838f8e0ac6eb4fb0e19d53c5de6f433425af4916fa9" gracePeriod=30 Dec 01 08:37:00 crc kubenswrapper[4744]: I1201 08:37:00.999091 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:37:00 crc kubenswrapper[4744]: I1201 08:37:00.999475 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="proxy-httpd" containerID="cri-o://d9e4f2a9231bd4415c72c8bc42a899e4d0cba0bf2dfd0f44aaeb71f401c9303b" gracePeriod=30 Dec 01 08:37:00 crc kubenswrapper[4744]: I1201 08:37:00.999574 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="sg-core" containerID="cri-o://9111e09f332b16dd57d30bf177e2e2096e7bb9b9a5c1cd20e2459af23f76740f" gracePeriod=30 Dec 01 08:37:00 crc kubenswrapper[4744]: I1201 08:37:00.999613 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-notification-agent" containerID="cri-o://15da7c51c0b82a6e91e1bc9596def580cddaa82df9899537a23d4933f357da05" gracePeriod=30 Dec 01 08:37:01 crc kubenswrapper[4744]: I1201 08:37:01.031842 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.161744674 podStartE2EDuration="13.031821879s" podCreationTimestamp="2025-12-01 08:36:48 +0000 UTC" firstStartedPulling="2025-12-01 08:36:48.907512354 +0000 UTC m=+1140.896570275" lastFinishedPulling="2025-12-01 08:36:59.777589559 +0000 UTC m=+1151.766647480" observedRunningTime="2025-12-01 08:37:01.02098176 +0000 UTC m=+1153.010039681" watchObservedRunningTime="2025-12-01 08:37:01.031821879 +0000 UTC m=+1153.020879800" Dec 01 08:37:02 crc kubenswrapper[4744]: I1201 08:37:02.013824 4744 generic.go:334] "Generic (PLEG): container finished" podID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerID="d9e4f2a9231bd4415c72c8bc42a899e4d0cba0bf2dfd0f44aaeb71f401c9303b" exitCode=0 Dec 01 08:37:02 crc kubenswrapper[4744]: I1201 08:37:02.014121 4744 generic.go:334] "Generic (PLEG): container finished" podID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerID="9111e09f332b16dd57d30bf177e2e2096e7bb9b9a5c1cd20e2459af23f76740f" exitCode=2 Dec 01 08:37:02 crc kubenswrapper[4744]: I1201 08:37:02.014131 4744 generic.go:334] "Generic (PLEG): container finished" podID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerID="ac755322f88c3b6773a34838f8e0ac6eb4fb0e19d53c5de6f433425af4916fa9" exitCode=0 Dec 01 08:37:02 crc kubenswrapper[4744]: I1201 08:37:02.013958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerDied","Data":"d9e4f2a9231bd4415c72c8bc42a899e4d0cba0bf2dfd0f44aaeb71f401c9303b"} Dec 01 08:37:02 crc kubenswrapper[4744]: I1201 08:37:02.014169 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerDied","Data":"9111e09f332b16dd57d30bf177e2e2096e7bb9b9a5c1cd20e2459af23f76740f"} Dec 01 08:37:02 crc kubenswrapper[4744]: I1201 08:37:02.014183 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerDied","Data":"ac755322f88c3b6773a34838f8e0ac6eb4fb0e19d53c5de6f433425af4916fa9"} Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.025704 4744 generic.go:334] "Generic (PLEG): container finished" podID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerID="15da7c51c0b82a6e91e1bc9596def580cddaa82df9899537a23d4933f357da05" exitCode=0 Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.025857 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerDied","Data":"15da7c51c0b82a6e91e1bc9596def580cddaa82df9899537a23d4933f357da05"} Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.151346 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.296372 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-combined-ca-bundle\") pod \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.296634 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-sg-core-conf-yaml\") pod \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.296722 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-scripts\") pod \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.296888 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-config-data\") pod \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.297049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-run-httpd\") pod \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.297215 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zkj9\" (UniqueName: \"kubernetes.io/projected/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-kube-api-access-5zkj9\") pod \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.297303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-log-httpd\") pod \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\" (UID: \"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c\") " Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.298052 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" (UID: "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.298635 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" (UID: "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.303112 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-scripts" (OuterVolumeSpecName: "scripts") pod "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" (UID: "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.323917 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-kube-api-access-5zkj9" (OuterVolumeSpecName: "kube-api-access-5zkj9") pod "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" (UID: "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c"). InnerVolumeSpecName "kube-api-access-5zkj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.339575 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" (UID: "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.383367 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" (UID: "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.399459 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.399493 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zkj9\" (UniqueName: \"kubernetes.io/projected/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-kube-api-access-5zkj9\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.399503 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.399512 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.399520 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.399527 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.412098 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-config-data" (OuterVolumeSpecName: "config-data") pod "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" (UID: "d5af0b31-72a4-4bb5-b2f6-8b0b307d695c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:03 crc kubenswrapper[4744]: I1201 08:37:03.500978 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.035852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d5af0b31-72a4-4bb5-b2f6-8b0b307d695c","Type":"ContainerDied","Data":"d1812ea927ee5251a8258bea5e90039cdd8d48c315e01fec0bd5243979d76070"} Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.036166 4744 scope.go:117] "RemoveContainer" containerID="d9e4f2a9231bd4415c72c8bc42a899e4d0cba0bf2dfd0f44aaeb71f401c9303b" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.036298 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.052596 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.053871 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.055310 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.055373 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5b7d595658-kpqqh" podUID="947af6f3-b9a2-46da-98d3-bed308f4ba29" containerName="heat-engine" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.074234 4744 scope.go:117] "RemoveContainer" containerID="9111e09f332b16dd57d30bf177e2e2096e7bb9b9a5c1cd20e2459af23f76740f" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.075016 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.086847 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.097906 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.098281 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerName="heat-api" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.098295 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerName="heat-api" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.098315 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="sg-core" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.098322 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="sg-core" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.098341 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="proxy-httpd" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.098350 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="proxy-httpd" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.098370 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-notification-agent" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.098377 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-notification-agent" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.098390 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-central-agent" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099167 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-central-agent" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.099197 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerName="heat-cfnapi" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099206 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerName="heat-cfnapi" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.099222 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerName="heat-cfnapi" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099230 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerName="heat-cfnapi" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099522 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerName="heat-cfnapi" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099540 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-central-agent" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099553 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9223dbb6-40f1-4946-b5e3-e84a1e25aa26" containerName="heat-cfnapi" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099567 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="ceilometer-notification-agent" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099579 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="sg-core" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099587 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerName="heat-api" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099597 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerName="heat-api" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099615 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" containerName="proxy-httpd" Dec 01 08:37:04 crc kubenswrapper[4744]: E1201 08:37:04.099830 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerName="heat-api" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.099840 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc1a384-3224-46e5-aadc-0421bb72626a" containerName="heat-api" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.102901 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.108753 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.109183 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.109698 4744 scope.go:117] "RemoveContainer" containerID="15da7c51c0b82a6e91e1bc9596def580cddaa82df9899537a23d4933f357da05" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.124646 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.151907 4744 scope.go:117] "RemoveContainer" containerID="ac755322f88c3b6773a34838f8e0ac6eb4fb0e19d53c5de6f433425af4916fa9" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.213525 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-log-httpd\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.213574 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbvgn\" (UniqueName: \"kubernetes.io/projected/dc4b9df2-9888-44d2-9ac2-acb53a516534-kube-api-access-vbvgn\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.213592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-run-httpd\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.213682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-config-data\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.213710 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-scripts\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.213728 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.213919 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.300093 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5af0b31-72a4-4bb5-b2f6-8b0b307d695c" path="/var/lib/kubelet/pods/d5af0b31-72a4-4bb5-b2f6-8b0b307d695c/volumes" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.320636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-config-data\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.320703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-scripts\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.320727 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.320795 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.320874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-log-httpd\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.322040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-log-httpd\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.322064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbvgn\" (UniqueName: \"kubernetes.io/projected/dc4b9df2-9888-44d2-9ac2-acb53a516534-kube-api-access-vbvgn\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.322530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-run-httpd\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.322886 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-run-httpd\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.326749 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-scripts\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.327516 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-config-data\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.328870 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.329444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.348748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbvgn\" (UniqueName: \"kubernetes.io/projected/dc4b9df2-9888-44d2-9ac2-acb53a516534-kube-api-access-vbvgn\") pod \"ceilometer-0\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.450668 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.896690 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:04 crc kubenswrapper[4744]: W1201 08:37:04.905192 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc4b9df2_9888_44d2_9ac2_acb53a516534.slice/crio-03f77889db447490fce6ffe67c5d654489c124deab37a58b6bf14bc45d89be7c WatchSource:0}: Error finding container 03f77889db447490fce6ffe67c5d654489c124deab37a58b6bf14bc45d89be7c: Status 404 returned error can't find the container with id 03f77889db447490fce6ffe67c5d654489c124deab37a58b6bf14bc45d89be7c Dec 01 08:37:04 crc kubenswrapper[4744]: I1201 08:37:04.987124 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.045994 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b7d595658-kpqqh" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.046374 4744 generic.go:334] "Generic (PLEG): container finished" podID="947af6f3-b9a2-46da-98d3-bed308f4ba29" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" exitCode=0 Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.046554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5b7d595658-kpqqh" event={"ID":"947af6f3-b9a2-46da-98d3-bed308f4ba29","Type":"ContainerDied","Data":"a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c"} Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.046607 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5b7d595658-kpqqh" event={"ID":"947af6f3-b9a2-46da-98d3-bed308f4ba29","Type":"ContainerDied","Data":"ee6e4ca6501a0f464e2478fcf8f0d647b92ff6c43b610d76cfa1011c67fe9758"} Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.046625 4744 scope.go:117] "RemoveContainer" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.050888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerStarted","Data":"03f77889db447490fce6ffe67c5d654489c124deab37a58b6bf14bc45d89be7c"} Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.075270 4744 scope.go:117] "RemoveContainer" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" Dec 01 08:37:05 crc kubenswrapper[4744]: E1201 08:37:05.076312 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c\": container with ID starting with a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c not found: ID does not exist" containerID="a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.076364 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c"} err="failed to get container status \"a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c\": rpc error: code = NotFound desc = could not find container \"a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c\": container with ID starting with a1c6be495eeb40ec57ae4c12daeaf6cb968d1af670fc5aabb45234061004fe2c not found: ID does not exist" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.142056 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-combined-ca-bundle\") pod \"947af6f3-b9a2-46da-98d3-bed308f4ba29\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.142166 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data\") pod \"947af6f3-b9a2-46da-98d3-bed308f4ba29\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.142250 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data-custom\") pod \"947af6f3-b9a2-46da-98d3-bed308f4ba29\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.142492 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2btk\" (UniqueName: \"kubernetes.io/projected/947af6f3-b9a2-46da-98d3-bed308f4ba29-kube-api-access-v2btk\") pod \"947af6f3-b9a2-46da-98d3-bed308f4ba29\" (UID: \"947af6f3-b9a2-46da-98d3-bed308f4ba29\") " Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.148696 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "947af6f3-b9a2-46da-98d3-bed308f4ba29" (UID: "947af6f3-b9a2-46da-98d3-bed308f4ba29"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.149752 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/947af6f3-b9a2-46da-98d3-bed308f4ba29-kube-api-access-v2btk" (OuterVolumeSpecName: "kube-api-access-v2btk") pod "947af6f3-b9a2-46da-98d3-bed308f4ba29" (UID: "947af6f3-b9a2-46da-98d3-bed308f4ba29"). InnerVolumeSpecName "kube-api-access-v2btk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.172796 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "947af6f3-b9a2-46da-98d3-bed308f4ba29" (UID: "947af6f3-b9a2-46da-98d3-bed308f4ba29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.209089 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data" (OuterVolumeSpecName: "config-data") pod "947af6f3-b9a2-46da-98d3-bed308f4ba29" (UID: "947af6f3-b9a2-46da-98d3-bed308f4ba29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.245201 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2btk\" (UniqueName: \"kubernetes.io/projected/947af6f3-b9a2-46da-98d3-bed308f4ba29-kube-api-access-v2btk\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.245231 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.245244 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.245256 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947af6f3-b9a2-46da-98d3-bed308f4ba29-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.391674 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5b7d595658-kpqqh"] Dec 01 08:37:05 crc kubenswrapper[4744]: I1201 08:37:05.400783 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-5b7d595658-kpqqh"] Dec 01 08:37:06 crc kubenswrapper[4744]: I1201 08:37:06.064388 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerStarted","Data":"149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9"} Dec 01 08:37:06 crc kubenswrapper[4744]: I1201 08:37:06.295161 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="947af6f3-b9a2-46da-98d3-bed308f4ba29" path="/var/lib/kubelet/pods/947af6f3-b9a2-46da-98d3-bed308f4ba29/volumes" Dec 01 08:37:07 crc kubenswrapper[4744]: I1201 08:37:07.078241 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerStarted","Data":"ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2"} Dec 01 08:37:07 crc kubenswrapper[4744]: I1201 08:37:07.078297 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerStarted","Data":"3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a"} Dec 01 08:37:09 crc kubenswrapper[4744]: I1201 08:37:09.100024 4744 generic.go:334] "Generic (PLEG): container finished" podID="522db2e5-3aed-45b0-a228-ab2b875a3cee" containerID="f12d78059710426c963a1cb1d2f3860cc3f000b0f18373f3a8687e7477a5cd6d" exitCode=0 Dec 01 08:37:09 crc kubenswrapper[4744]: I1201 08:37:09.100393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t86st" event={"ID":"522db2e5-3aed-45b0-a228-ab2b875a3cee","Type":"ContainerDied","Data":"f12d78059710426c963a1cb1d2f3860cc3f000b0f18373f3a8687e7477a5cd6d"} Dec 01 08:37:09 crc kubenswrapper[4744]: I1201 08:37:09.105204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerStarted","Data":"abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9"} Dec 01 08:37:09 crc kubenswrapper[4744]: I1201 08:37:09.106342 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:37:09 crc kubenswrapper[4744]: I1201 08:37:09.143502 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.853196074 podStartE2EDuration="5.143488696s" podCreationTimestamp="2025-12-01 08:37:04 +0000 UTC" firstStartedPulling="2025-12-01 08:37:04.908554958 +0000 UTC m=+1156.897612879" lastFinishedPulling="2025-12-01 08:37:08.19884754 +0000 UTC m=+1160.187905501" observedRunningTime="2025-12-01 08:37:09.13939329 +0000 UTC m=+1161.128451211" watchObservedRunningTime="2025-12-01 08:37:09.143488696 +0000 UTC m=+1161.132546617" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.471332 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.665314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-config-data\") pod \"522db2e5-3aed-45b0-a228-ab2b875a3cee\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.665552 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w2wl\" (UniqueName: \"kubernetes.io/projected/522db2e5-3aed-45b0-a228-ab2b875a3cee-kube-api-access-5w2wl\") pod \"522db2e5-3aed-45b0-a228-ab2b875a3cee\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.665591 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-scripts\") pod \"522db2e5-3aed-45b0-a228-ab2b875a3cee\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.665722 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-combined-ca-bundle\") pod \"522db2e5-3aed-45b0-a228-ab2b875a3cee\" (UID: \"522db2e5-3aed-45b0-a228-ab2b875a3cee\") " Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.672578 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522db2e5-3aed-45b0-a228-ab2b875a3cee-kube-api-access-5w2wl" (OuterVolumeSpecName: "kube-api-access-5w2wl") pod "522db2e5-3aed-45b0-a228-ab2b875a3cee" (UID: "522db2e5-3aed-45b0-a228-ab2b875a3cee"). InnerVolumeSpecName "kube-api-access-5w2wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.672789 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-scripts" (OuterVolumeSpecName: "scripts") pod "522db2e5-3aed-45b0-a228-ab2b875a3cee" (UID: "522db2e5-3aed-45b0-a228-ab2b875a3cee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.700774 4744 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod1691c44a-f2ba-43eb-9ff3-a6a25b352d78"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod1691c44a-f2ba-43eb-9ff3-a6a25b352d78] : Timed out while waiting for systemd to remove kubepods-besteffort-pod1691c44a_f2ba_43eb_9ff3_a6a25b352d78.slice" Dec 01 08:37:10 crc kubenswrapper[4744]: E1201 08:37:10.700838 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod1691c44a-f2ba-43eb-9ff3-a6a25b352d78] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod1691c44a-f2ba-43eb-9ff3-a6a25b352d78] : Timed out while waiting for systemd to remove kubepods-besteffort-pod1691c44a_f2ba_43eb_9ff3_a6a25b352d78.slice" pod="openstack/glance-default-external-api-0" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.704847 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "522db2e5-3aed-45b0-a228-ab2b875a3cee" (UID: "522db2e5-3aed-45b0-a228-ab2b875a3cee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.740520 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-config-data" (OuterVolumeSpecName: "config-data") pod "522db2e5-3aed-45b0-a228-ab2b875a3cee" (UID: "522db2e5-3aed-45b0-a228-ab2b875a3cee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.768858 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w2wl\" (UniqueName: \"kubernetes.io/projected/522db2e5-3aed-45b0-a228-ab2b875a3cee-kube-api-access-5w2wl\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.768900 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.768915 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:10 crc kubenswrapper[4744]: I1201 08:37:10.768927 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522db2e5-3aed-45b0-a228-ab2b875a3cee-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.124174 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.124168 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-t86st" event={"ID":"522db2e5-3aed-45b0-a228-ab2b875a3cee","Type":"ContainerDied","Data":"9efe2ac4dbde65e1061d55e8d1e051b23e3ffd47403fdbf4101cff3264afe796"} Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.124227 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9efe2ac4dbde65e1061d55e8d1e051b23e3ffd47403fdbf4101cff3264afe796" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.124624 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-t86st" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.176020 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.214243 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.227377 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:37:11 crc kubenswrapper[4744]: E1201 08:37:11.228010 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947af6f3-b9a2-46da-98d3-bed308f4ba29" containerName="heat-engine" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.228142 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="947af6f3-b9a2-46da-98d3-bed308f4ba29" containerName="heat-engine" Dec 01 08:37:11 crc kubenswrapper[4744]: E1201 08:37:11.228266 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522db2e5-3aed-45b0-a228-ab2b875a3cee" containerName="nova-cell0-conductor-db-sync" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.228366 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="522db2e5-3aed-45b0-a228-ab2b875a3cee" containerName="nova-cell0-conductor-db-sync" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.228770 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="522db2e5-3aed-45b0-a228-ab2b875a3cee" containerName="nova-cell0-conductor-db-sync" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.228888 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="947af6f3-b9a2-46da-98d3-bed308f4ba29" containerName="heat-engine" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.230174 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.234320 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.234607 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.243326 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.265837 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.267167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.270850 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s85xn" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.271300 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.279280 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.379677 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52e8d56-63ad-466d-a7e9-57c75c346a1c-logs\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.379735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqk44\" (UniqueName: \"kubernetes.io/projected/a52e8d56-63ad-466d-a7e9-57c75c346a1c-kube-api-access-vqk44\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.379802 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ab1910-715b-4982-9903-453336b14b78-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.379844 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.379946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.380039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a52e8d56-63ad-466d-a7e9-57c75c346a1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.380133 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.380246 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.380269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ab1910-715b-4982-9903-453336b14b78-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.380878 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjvdf\" (UniqueName: \"kubernetes.io/projected/d4ab1910-715b-4982-9903-453336b14b78-kube-api-access-qjvdf\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.380966 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482279 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a52e8d56-63ad-466d-a7e9-57c75c346a1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482373 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482441 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482463 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ab1910-715b-4982-9903-453336b14b78-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482532 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjvdf\" (UniqueName: \"kubernetes.io/projected/d4ab1910-715b-4982-9903-453336b14b78-kube-api-access-qjvdf\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482550 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482576 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52e8d56-63ad-466d-a7e9-57c75c346a1c-logs\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqk44\" (UniqueName: \"kubernetes.io/projected/a52e8d56-63ad-466d-a7e9-57c75c346a1c-kube-api-access-vqk44\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ab1910-715b-4982-9903-453336b14b78-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.482661 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.483199 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.483557 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52e8d56-63ad-466d-a7e9-57c75c346a1c-logs\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.491700 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.491989 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.492789 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4ab1910-715b-4982-9903-453336b14b78-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.492949 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4ab1910-715b-4982-9903-453336b14b78-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.493392 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.495105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52e8d56-63ad-466d-a7e9-57c75c346a1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.495665 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a52e8d56-63ad-466d-a7e9-57c75c346a1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.508904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjvdf\" (UniqueName: \"kubernetes.io/projected/d4ab1910-715b-4982-9903-453336b14b78-kube-api-access-qjvdf\") pod \"nova-cell0-conductor-0\" (UID: \"d4ab1910-715b-4982-9903-453336b14b78\") " pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.523973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqk44\" (UniqueName: \"kubernetes.io/projected/a52e8d56-63ad-466d-a7e9-57c75c346a1c-kube-api-access-vqk44\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.526462 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"a52e8d56-63ad-466d-a7e9-57c75c346a1c\") " pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.560844 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 08:37:11 crc kubenswrapper[4744]: I1201 08:37:11.590999 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:12 crc kubenswrapper[4744]: W1201 08:37:12.102422 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda52e8d56_63ad_466d_a7e9_57c75c346a1c.slice/crio-27b637624abdc291e647d82fb50735ee0b0d63081e97e9b67413050eb3dbdbae WatchSource:0}: Error finding container 27b637624abdc291e647d82fb50735ee0b0d63081e97e9b67413050eb3dbdbae: Status 404 returned error can't find the container with id 27b637624abdc291e647d82fb50735ee0b0d63081e97e9b67413050eb3dbdbae Dec 01 08:37:12 crc kubenswrapper[4744]: I1201 08:37:12.104628 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 08:37:12 crc kubenswrapper[4744]: I1201 08:37:12.121753 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 08:37:12 crc kubenswrapper[4744]: I1201 08:37:12.145895 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a52e8d56-63ad-466d-a7e9-57c75c346a1c","Type":"ContainerStarted","Data":"27b637624abdc291e647d82fb50735ee0b0d63081e97e9b67413050eb3dbdbae"} Dec 01 08:37:12 crc kubenswrapper[4744]: I1201 08:37:12.299283 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1691c44a-f2ba-43eb-9ff3-a6a25b352d78" path="/var/lib/kubelet/pods/1691c44a-f2ba-43eb-9ff3-a6a25b352d78/volumes" Dec 01 08:37:13 crc kubenswrapper[4744]: I1201 08:37:13.196612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a52e8d56-63ad-466d-a7e9-57c75c346a1c","Type":"ContainerStarted","Data":"4967241f81a88884358c373670c5f48951d3c3b60cb7471408915e441ea2a63e"} Dec 01 08:37:13 crc kubenswrapper[4744]: I1201 08:37:13.244068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d4ab1910-715b-4982-9903-453336b14b78","Type":"ContainerStarted","Data":"54a883a83de66f300fbd7f485d82d5407fa8627d6fc0cb08afea24760753baec"} Dec 01 08:37:13 crc kubenswrapper[4744]: I1201 08:37:13.244471 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d4ab1910-715b-4982-9903-453336b14b78","Type":"ContainerStarted","Data":"11202d083ba7e05c8b39b7709eb3209f0252797d30bce0eb2b9ff01233e79abc"} Dec 01 08:37:13 crc kubenswrapper[4744]: I1201 08:37:13.245710 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:13 crc kubenswrapper[4744]: I1201 08:37:13.306936 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.306916839 podStartE2EDuration="2.306916839s" podCreationTimestamp="2025-12-01 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:13.291684275 +0000 UTC m=+1165.280742196" watchObservedRunningTime="2025-12-01 08:37:13.306916839 +0000 UTC m=+1165.295974760" Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.102485 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.102801 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-central-agent" containerID="cri-o://149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9" gracePeriod=30 Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.102896 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="sg-core" containerID="cri-o://ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2" gracePeriod=30 Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.102867 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="proxy-httpd" containerID="cri-o://abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9" gracePeriod=30 Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.102863 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-notification-agent" containerID="cri-o://3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a" gracePeriod=30 Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.259108 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a52e8d56-63ad-466d-a7e9-57c75c346a1c","Type":"ContainerStarted","Data":"24631348e19d29581aeb1c837c5c67653ea058a526b4739e1acae57e593a5369"} Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.268387 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerID="abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9" exitCode=0 Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.268451 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerID="ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2" exitCode=2 Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.269600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerDied","Data":"abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9"} Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.269636 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerDied","Data":"ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2"} Dec 01 08:37:14 crc kubenswrapper[4744]: I1201 08:37:14.287093 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.287075676 podStartE2EDuration="3.287075676s" podCreationTimestamp="2025-12-01 08:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:14.28229061 +0000 UTC m=+1166.271348541" watchObservedRunningTime="2025-12-01 08:37:14.287075676 +0000 UTC m=+1166.276133597" Dec 01 08:37:15 crc kubenswrapper[4744]: I1201 08:37:15.279712 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerID="3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a" exitCode=0 Dec 01 08:37:15 crc kubenswrapper[4744]: I1201 08:37:15.280474 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerDied","Data":"3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a"} Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.058646 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.168568 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-combined-ca-bundle\") pod \"dc4b9df2-9888-44d2-9ac2-acb53a516534\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.168646 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-sg-core-conf-yaml\") pod \"dc4b9df2-9888-44d2-9ac2-acb53a516534\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.168746 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-run-httpd\") pod \"dc4b9df2-9888-44d2-9ac2-acb53a516534\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.168773 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-log-httpd\") pod \"dc4b9df2-9888-44d2-9ac2-acb53a516534\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.168810 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-config-data\") pod \"dc4b9df2-9888-44d2-9ac2-acb53a516534\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.168843 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbvgn\" (UniqueName: \"kubernetes.io/projected/dc4b9df2-9888-44d2-9ac2-acb53a516534-kube-api-access-vbvgn\") pod \"dc4b9df2-9888-44d2-9ac2-acb53a516534\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.168912 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-scripts\") pod \"dc4b9df2-9888-44d2-9ac2-acb53a516534\" (UID: \"dc4b9df2-9888-44d2-9ac2-acb53a516534\") " Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.169474 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dc4b9df2-9888-44d2-9ac2-acb53a516534" (UID: "dc4b9df2-9888-44d2-9ac2-acb53a516534"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.169672 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dc4b9df2-9888-44d2-9ac2-acb53a516534" (UID: "dc4b9df2-9888-44d2-9ac2-acb53a516534"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.174401 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-scripts" (OuterVolumeSpecName: "scripts") pod "dc4b9df2-9888-44d2-9ac2-acb53a516534" (UID: "dc4b9df2-9888-44d2-9ac2-acb53a516534"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.180569 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4b9df2-9888-44d2-9ac2-acb53a516534-kube-api-access-vbvgn" (OuterVolumeSpecName: "kube-api-access-vbvgn") pod "dc4b9df2-9888-44d2-9ac2-acb53a516534" (UID: "dc4b9df2-9888-44d2-9ac2-acb53a516534"). InnerVolumeSpecName "kube-api-access-vbvgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.204628 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dc4b9df2-9888-44d2-9ac2-acb53a516534" (UID: "dc4b9df2-9888-44d2-9ac2-acb53a516534"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.272023 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.273332 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbvgn\" (UniqueName: \"kubernetes.io/projected/dc4b9df2-9888-44d2-9ac2-acb53a516534-kube-api-access-vbvgn\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.273914 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.274533 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.274631 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc4b9df2-9888-44d2-9ac2-acb53a516534-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.277579 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc4b9df2-9888-44d2-9ac2-acb53a516534" (UID: "dc4b9df2-9888-44d2-9ac2-acb53a516534"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.291876 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerID="149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9" exitCode=0 Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.291942 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.301590 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerDied","Data":"149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9"} Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.301648 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc4b9df2-9888-44d2-9ac2-acb53a516534","Type":"ContainerDied","Data":"03f77889db447490fce6ffe67c5d654489c124deab37a58b6bf14bc45d89be7c"} Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.301673 4744 scope.go:117] "RemoveContainer" containerID="abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.328723 4744 scope.go:117] "RemoveContainer" containerID="ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.350800 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-config-data" (OuterVolumeSpecName: "config-data") pod "dc4b9df2-9888-44d2-9ac2-acb53a516534" (UID: "dc4b9df2-9888-44d2-9ac2-acb53a516534"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.351671 4744 scope.go:117] "RemoveContainer" containerID="3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.373223 4744 scope.go:117] "RemoveContainer" containerID="149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.387832 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.387862 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc4b9df2-9888-44d2-9ac2-acb53a516534-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.395836 4744 scope.go:117] "RemoveContainer" containerID="abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9" Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.396453 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9\": container with ID starting with abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9 not found: ID does not exist" containerID="abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.396491 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9"} err="failed to get container status \"abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9\": rpc error: code = NotFound desc = could not find container \"abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9\": container with ID starting with abc448a7fd29d147f557d90fb2b4d12c5b613917add0e32773067bf34fb601f9 not found: ID does not exist" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.396523 4744 scope.go:117] "RemoveContainer" containerID="ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2" Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.396801 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2\": container with ID starting with ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2 not found: ID does not exist" containerID="ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.396823 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2"} err="failed to get container status \"ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2\": rpc error: code = NotFound desc = could not find container \"ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2\": container with ID starting with ab8685894bce05dcd61efa42f2b92fb3f8c1a30d93fcc8fab8f806ef242dccf2 not found: ID does not exist" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.396840 4744 scope.go:117] "RemoveContainer" containerID="3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a" Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.397144 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a\": container with ID starting with 3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a not found: ID does not exist" containerID="3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.397169 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a"} err="failed to get container status \"3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a\": rpc error: code = NotFound desc = could not find container \"3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a\": container with ID starting with 3529ce740ff093b371bf85be878da95bc2fd8e2a7be998dd721df59827d9cb7a not found: ID does not exist" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.397186 4744 scope.go:117] "RemoveContainer" containerID="149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9" Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.397380 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9\": container with ID starting with 149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9 not found: ID does not exist" containerID="149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.397400 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9"} err="failed to get container status \"149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9\": rpc error: code = NotFound desc = could not find container \"149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9\": container with ID starting with 149bd2878d1d7bb92eae5c19145a327a94890b59ba01b0fa1ff25e52c522c0c9 not found: ID does not exist" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.662734 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.674400 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.689476 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.690974 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-notification-agent" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691003 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-notification-agent" Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.691037 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="proxy-httpd" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691048 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="proxy-httpd" Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.691098 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-central-agent" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691110 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-central-agent" Dec 01 08:37:16 crc kubenswrapper[4744]: E1201 08:37:16.691126 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="sg-core" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691137 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="sg-core" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691464 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-notification-agent" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691503 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="proxy-httpd" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691528 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="ceilometer-central-agent" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.691546 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" containerName="sg-core" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.694385 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.696931 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.697134 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.714322 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.800509 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.800585 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-config-data\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.800631 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.800668 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-run-httpd\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.800702 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncdcc\" (UniqueName: \"kubernetes.io/projected/5ef069dd-c0d2-4235-b07c-e8ea6798d661-kube-api-access-ncdcc\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.800751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-scripts\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.800808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-log-httpd\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.902700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.902779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-run-httpd\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.902820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncdcc\" (UniqueName: \"kubernetes.io/projected/5ef069dd-c0d2-4235-b07c-e8ea6798d661-kube-api-access-ncdcc\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.902876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-scripts\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.902946 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-log-httpd\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.902996 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.903058 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-config-data\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.904473 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-log-httpd\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.904592 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-run-httpd\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.908765 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-config-data\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.909327 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-scripts\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.910463 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.914505 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:16 crc kubenswrapper[4744]: I1201 08:37:16.919162 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncdcc\" (UniqueName: \"kubernetes.io/projected/5ef069dd-c0d2-4235-b07c-e8ea6798d661-kube-api-access-ncdcc\") pod \"ceilometer-0\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " pod="openstack/ceilometer-0" Dec 01 08:37:17 crc kubenswrapper[4744]: I1201 08:37:17.029343 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:17 crc kubenswrapper[4744]: I1201 08:37:17.483300 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:18 crc kubenswrapper[4744]: I1201 08:37:18.299124 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc4b9df2-9888-44d2-9ac2-acb53a516534" path="/var/lib/kubelet/pods/dc4b9df2-9888-44d2-9ac2-acb53a516534/volumes" Dec 01 08:37:18 crc kubenswrapper[4744]: I1201 08:37:18.312548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerStarted","Data":"53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0"} Dec 01 08:37:18 crc kubenswrapper[4744]: I1201 08:37:18.312613 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerStarted","Data":"43e4f5fd0801cc7d0ee852f6a879832f52d67ad5f932cf8343af7d003fbf9395"} Dec 01 08:37:18 crc kubenswrapper[4744]: I1201 08:37:18.635185 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:37:18 crc kubenswrapper[4744]: I1201 08:37:18.635329 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:37:19 crc kubenswrapper[4744]: I1201 08:37:19.327108 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerStarted","Data":"1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d"} Dec 01 08:37:20 crc kubenswrapper[4744]: I1201 08:37:20.342870 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerStarted","Data":"2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b"} Dec 01 08:37:21 crc kubenswrapper[4744]: I1201 08:37:21.561612 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 08:37:21 crc kubenswrapper[4744]: I1201 08:37:21.562401 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 08:37:21 crc kubenswrapper[4744]: I1201 08:37:21.598279 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 08:37:21 crc kubenswrapper[4744]: I1201 08:37:21.628832 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 08:37:21 crc kubenswrapper[4744]: I1201 08:37:21.632017 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.132562 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-jc998"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.134295 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.136554 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.140842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.153254 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jc998"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.300595 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpfcv\" (UniqueName: \"kubernetes.io/projected/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-kube-api-access-mpfcv\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.300682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-scripts\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.300807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-config-data\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.301090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.348512 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.350113 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.357291 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.371876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerStarted","Data":"c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae"} Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.371916 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.371929 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.371938 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.406359 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-config-data\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.406647 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.407680 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpfcv\" (UniqueName: \"kubernetes.io/projected/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-kube-api-access-mpfcv\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.407837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-scripts\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.414002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-scripts\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.414165 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-config-data\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.459802 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.510288 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-config-data\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.510419 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f72bbad-2cab-421b-9f3c-2cd27581c638-logs\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.510520 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dh7p\" (UniqueName: \"kubernetes.io/projected/1f72bbad-2cab-421b-9f3c-2cd27581c638-kube-api-access-4dh7p\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.510560 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.523353 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpfcv\" (UniqueName: \"kubernetes.io/projected/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-kube-api-access-mpfcv\") pod \"nova-cell0-cell-mapping-jc998\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.553476 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.565570 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.771242245 podStartE2EDuration="6.565551479s" podCreationTimestamp="2025-12-01 08:37:16 +0000 UTC" firstStartedPulling="2025-12-01 08:37:17.489175158 +0000 UTC m=+1169.478233079" lastFinishedPulling="2025-12-01 08:37:21.283484382 +0000 UTC m=+1173.272542313" observedRunningTime="2025-12-01 08:37:22.409289761 +0000 UTC m=+1174.398347682" watchObservedRunningTime="2025-12-01 08:37:22.565551479 +0000 UTC m=+1174.554609400" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.585465 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.587019 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.595192 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.611162 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.612929 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-config-data\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.613027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f72bbad-2cab-421b-9f3c-2cd27581c638-logs\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.613105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dh7p\" (UniqueName: \"kubernetes.io/projected/1f72bbad-2cab-421b-9f3c-2cd27581c638-kube-api-access-4dh7p\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.613149 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.615099 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f72bbad-2cab-421b-9f3c-2cd27581c638-logs\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.617836 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.619086 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.621010 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.621042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-config-data\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.626462 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7877d89589-cm67w"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.628153 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.628818 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.632712 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.635986 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.646467 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.646821 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.655998 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dh7p\" (UniqueName: \"kubernetes.io/projected/1f72bbad-2cab-421b-9f3c-2cd27581c638-kube-api-access-4dh7p\") pod \"nova-metadata-0\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.672849 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.673373 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-cm67w"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.697513 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.716914 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.717196 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-config-data\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.717383 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.717513 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtnvt\" (UniqueName: \"kubernetes.io/projected/ce3cbc79-3564-42d0-9005-a9c503c25f00-kube-api-access-rtnvt\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.717958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w59bl\" (UniqueName: \"kubernetes.io/projected/92d02f35-1811-4824-a1c8-375ebde20065-kube-api-access-w59bl\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718109 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-config-data\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718275 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3cbc79-3564-42d0-9005-a9c503c25f00-logs\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718394 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718496 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718574 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsjdd\" (UniqueName: \"kubernetes.io/projected/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-kube-api-access-dsjdd\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718660 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-config\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.718767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-svc\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.758653 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.821910 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-config\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822194 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-svc\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-config-data\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822443 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtnvt\" (UniqueName: \"kubernetes.io/projected/ce3cbc79-3564-42d0-9005-a9c503c25f00-kube-api-access-rtnvt\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w59bl\" (UniqueName: \"kubernetes.io/projected/92d02f35-1811-4824-a1c8-375ebde20065-kube-api-access-w59bl\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822511 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822532 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-config-data\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3cbc79-3564-42d0-9005-a9c503c25f00-logs\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822612 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twmg8\" (UniqueName: \"kubernetes.io/projected/90e0478c-e863-4c2f-852c-673d676868c2-kube-api-access-twmg8\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822647 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.822706 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsjdd\" (UniqueName: \"kubernetes.io/projected/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-kube-api-access-dsjdd\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.823845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3cbc79-3564-42d0-9005-a9c503c25f00-logs\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.827499 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-config\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.831194 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-config-data\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.832358 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.833708 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-config-data\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.838187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.838307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.840171 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.849148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.849738 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-svc\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.863037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtnvt\" (UniqueName: \"kubernetes.io/projected/ce3cbc79-3564-42d0-9005-a9c503c25f00-kube-api-access-rtnvt\") pod \"nova-api-0\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.863203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsjdd\" (UniqueName: \"kubernetes.io/projected/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-kube-api-access-dsjdd\") pod \"dnsmasq-dns-7877d89589-cm67w\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.863479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w59bl\" (UniqueName: \"kubernetes.io/projected/92d02f35-1811-4824-a1c8-375ebde20065-kube-api-access-w59bl\") pod \"nova-scheduler-0\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.915809 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.924585 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.924630 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twmg8\" (UniqueName: \"kubernetes.io/projected/90e0478c-e863-4c2f-852c-673d676868c2-kube-api-access-twmg8\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.924767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.941104 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.944126 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twmg8\" (UniqueName: \"kubernetes.io/projected/90e0478c-e863-4c2f-852c-673d676868c2-kube-api-access-twmg8\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:22 crc kubenswrapper[4744]: I1201 08:37:22.945912 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.057838 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.070468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.086975 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.201631 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.276652 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.308383 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-24cgv"] Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.309776 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.312839 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.315574 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.320121 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-24cgv"] Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.360248 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jc998"] Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.411207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce3cbc79-3564-42d0-9005-a9c503c25f00","Type":"ContainerStarted","Data":"68831277585c33f25a037fa64b6a32238dfc9c21c04388a2383d6fd69d87e604"} Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.427746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f72bbad-2cab-421b-9f3c-2cd27581c638","Type":"ContainerStarted","Data":"8a6f1d14c2de2289913988be9fd35f2d3b833bae727c52c3c6b9685ad50c5eeb"} Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.436939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.437018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-scripts\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.437058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-config-data\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.437161 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh4kl\" (UniqueName: \"kubernetes.io/projected/3eb87b97-a941-413c-9778-29a593d44ef0-kube-api-access-jh4kl\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.538720 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-config-data\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.538804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh4kl\" (UniqueName: \"kubernetes.io/projected/3eb87b97-a941-413c-9778-29a593d44ef0-kube-api-access-jh4kl\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.538915 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.538977 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-scripts\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.548805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-scripts\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.554304 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.557937 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh4kl\" (UniqueName: \"kubernetes.io/projected/3eb87b97-a941-413c-9778-29a593d44ef0-kube-api-access-jh4kl\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.560246 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-config-data\") pod \"nova-cell1-conductor-db-sync-24cgv\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.636456 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:23 crc kubenswrapper[4744]: W1201 08:37:23.639908 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5129f5b1_84f9_4c6c_95ec_c0a8396ac082.slice/crio-8a27b689c67fe164c070c4a7a662b88d83b3172a829112bc783505ab33d7cf33 WatchSource:0}: Error finding container 8a27b689c67fe164c070c4a7a662b88d83b3172a829112bc783505ab33d7cf33: Status 404 returned error can't find the container with id 8a27b689c67fe164c070c4a7a662b88d83b3172a829112bc783505ab33d7cf33 Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.641707 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-cm67w"] Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.645823 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.767183 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:23 crc kubenswrapper[4744]: W1201 08:37:23.775035 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90e0478c_e863_4c2f_852c_673d676868c2.slice/crio-ccedb8b640c94510004f266c18a79c55269ca8f36af32aee9e6b6878c450b1fe WatchSource:0}: Error finding container ccedb8b640c94510004f266c18a79c55269ca8f36af32aee9e6b6878c450b1fe: Status 404 returned error can't find the container with id ccedb8b640c94510004f266c18a79c55269ca8f36af32aee9e6b6878c450b1fe Dec 01 08:37:23 crc kubenswrapper[4744]: I1201 08:37:23.914861 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-24cgv"] Dec 01 08:37:23 crc kubenswrapper[4744]: W1201 08:37:23.920427 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3eb87b97_a941_413c_9778_29a593d44ef0.slice/crio-d900dbc4a27a04d16bafade8317a9e12d856187b9d3ee0f6390194e5433d9cb1 WatchSource:0}: Error finding container d900dbc4a27a04d16bafade8317a9e12d856187b9d3ee0f6390194e5433d9cb1: Status 404 returned error can't find the container with id d900dbc4a27a04d16bafade8317a9e12d856187b9d3ee0f6390194e5433d9cb1 Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.462543 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-24cgv" event={"ID":"3eb87b97-a941-413c-9778-29a593d44ef0","Type":"ContainerStarted","Data":"b2a2a496c6333eb838de95bcb18cdb525d89b52ecaaeee3af278068c84bf73bf"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.462800 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-24cgv" event={"ID":"3eb87b97-a941-413c-9778-29a593d44ef0","Type":"ContainerStarted","Data":"d900dbc4a27a04d16bafade8317a9e12d856187b9d3ee0f6390194e5433d9cb1"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.466386 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92d02f35-1811-4824-a1c8-375ebde20065","Type":"ContainerStarted","Data":"8330e6a909b874935025c0442c9ff5a31614484702eaf1eba6a3c86b26be050d"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.470998 4744 generic.go:334] "Generic (PLEG): container finished" podID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerID="247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0" exitCode=0 Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.471079 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-cm67w" event={"ID":"5129f5b1-84f9-4c6c-95ec-c0a8396ac082","Type":"ContainerDied","Data":"247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.471117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-cm67w" event={"ID":"5129f5b1-84f9-4c6c-95ec-c0a8396ac082","Type":"ContainerStarted","Data":"8a27b689c67fe164c070c4a7a662b88d83b3172a829112bc783505ab33d7cf33"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.473491 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jc998" event={"ID":"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf","Type":"ContainerStarted","Data":"789cde6612119cbc9df1a36a57e6970528223e416fbfee18e1cb1cfe62d93173"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.473526 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jc998" event={"ID":"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf","Type":"ContainerStarted","Data":"0b0bff0e0f14a5233cc609c038fdef1be5df5298cbb614cde9679155ed73b37a"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.477394 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.477430 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.478221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"90e0478c-e863-4c2f-852c-673d676868c2","Type":"ContainerStarted","Data":"ccedb8b640c94510004f266c18a79c55269ca8f36af32aee9e6b6878c450b1fe"} Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.484327 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-24cgv" podStartSLOduration=1.484313003 podStartE2EDuration="1.484313003s" podCreationTimestamp="2025-12-01 08:37:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:24.480085772 +0000 UTC m=+1176.469143693" watchObservedRunningTime="2025-12-01 08:37:24.484313003 +0000 UTC m=+1176.473370924" Dec 01 08:37:24 crc kubenswrapper[4744]: I1201 08:37:24.547493 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-jc998" podStartSLOduration=2.547469225 podStartE2EDuration="2.547469225s" podCreationTimestamp="2025-12-01 08:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:24.538715325 +0000 UTC m=+1176.527773266" watchObservedRunningTime="2025-12-01 08:37:24.547469225 +0000 UTC m=+1176.536527146" Dec 01 08:37:25 crc kubenswrapper[4744]: I1201 08:37:25.430818 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 08:37:25 crc kubenswrapper[4744]: I1201 08:37:25.433363 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 08:37:25 crc kubenswrapper[4744]: I1201 08:37:25.497963 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-cm67w" event={"ID":"5129f5b1-84f9-4c6c-95ec-c0a8396ac082","Type":"ContainerStarted","Data":"f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2"} Dec 01 08:37:25 crc kubenswrapper[4744]: I1201 08:37:25.498061 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:25 crc kubenswrapper[4744]: I1201 08:37:25.520694 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7877d89589-cm67w" podStartSLOduration=3.52067963 podStartE2EDuration="3.52067963s" podCreationTimestamp="2025-12-01 08:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:25.518668353 +0000 UTC m=+1177.507726274" watchObservedRunningTime="2025-12-01 08:37:25.52067963 +0000 UTC m=+1177.509737551" Dec 01 08:37:26 crc kubenswrapper[4744]: I1201 08:37:26.049077 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:26 crc kubenswrapper[4744]: I1201 08:37:26.062546 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.524622 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"90e0478c-e863-4c2f-852c-673d676868c2","Type":"ContainerStarted","Data":"b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4"} Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.524670 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="90e0478c-e863-4c2f-852c-673d676868c2" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4" gracePeriod=30 Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.526471 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f72bbad-2cab-421b-9f3c-2cd27581c638","Type":"ContainerStarted","Data":"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4"} Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.526524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f72bbad-2cab-421b-9f3c-2cd27581c638","Type":"ContainerStarted","Data":"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559"} Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.526616 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-log" containerID="cri-o://75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559" gracePeriod=30 Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.526658 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-metadata" containerID="cri-o://922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4" gracePeriod=30 Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.529143 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92d02f35-1811-4824-a1c8-375ebde20065","Type":"ContainerStarted","Data":"a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e"} Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.531709 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce3cbc79-3564-42d0-9005-a9c503c25f00","Type":"ContainerStarted","Data":"853334c399072bc499caa4ab92ef6484754bcfcdf9ad564e62a5b3663f302364"} Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.531747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce3cbc79-3564-42d0-9005-a9c503c25f00","Type":"ContainerStarted","Data":"2557df3e2a35ca8c183f221af58573cea9a46f7dbcb7ebcf551e0a308534a886"} Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.573727 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.240705343 podStartE2EDuration="6.573702195s" podCreationTimestamp="2025-12-01 08:37:22 +0000 UTC" firstStartedPulling="2025-12-01 08:37:23.223911423 +0000 UTC m=+1175.212969344" lastFinishedPulling="2025-12-01 08:37:27.556908235 +0000 UTC m=+1179.545966196" observedRunningTime="2025-12-01 08:37:28.571299656 +0000 UTC m=+1180.560357587" watchObservedRunningTime="2025-12-01 08:37:28.573702195 +0000 UTC m=+1180.562760126" Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.586241 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.832896049 podStartE2EDuration="6.586209372s" podCreationTimestamp="2025-12-01 08:37:22 +0000 UTC" firstStartedPulling="2025-12-01 08:37:23.798526547 +0000 UTC m=+1175.787584468" lastFinishedPulling="2025-12-01 08:37:27.55183987 +0000 UTC m=+1179.540897791" observedRunningTime="2025-12-01 08:37:28.55670317 +0000 UTC m=+1180.545761111" watchObservedRunningTime="2025-12-01 08:37:28.586209372 +0000 UTC m=+1180.575267303" Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.622006 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.708234911 podStartE2EDuration="6.621967782s" podCreationTimestamp="2025-12-01 08:37:22 +0000 UTC" firstStartedPulling="2025-12-01 08:37:23.64161778 +0000 UTC m=+1175.630675701" lastFinishedPulling="2025-12-01 08:37:27.555350651 +0000 UTC m=+1179.544408572" observedRunningTime="2025-12-01 08:37:28.615863798 +0000 UTC m=+1180.604921729" watchObservedRunningTime="2025-12-01 08:37:28.621967782 +0000 UTC m=+1180.611025713" Dec 01 08:37:28 crc kubenswrapper[4744]: I1201 08:37:28.627115 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.355737694 podStartE2EDuration="6.627102878s" podCreationTimestamp="2025-12-01 08:37:22 +0000 UTC" firstStartedPulling="2025-12-01 08:37:23.278980124 +0000 UTC m=+1175.268038035" lastFinishedPulling="2025-12-01 08:37:27.550345298 +0000 UTC m=+1179.539403219" observedRunningTime="2025-12-01 08:37:28.601541419 +0000 UTC m=+1180.590599340" watchObservedRunningTime="2025-12-01 08:37:28.627102878 +0000 UTC m=+1180.616160809" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.106487 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.271242 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f72bbad-2cab-421b-9f3c-2cd27581c638-logs\") pod \"1f72bbad-2cab-421b-9f3c-2cd27581c638\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.271450 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-config-data\") pod \"1f72bbad-2cab-421b-9f3c-2cd27581c638\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.271500 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-combined-ca-bundle\") pod \"1f72bbad-2cab-421b-9f3c-2cd27581c638\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.271567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dh7p\" (UniqueName: \"kubernetes.io/projected/1f72bbad-2cab-421b-9f3c-2cd27581c638-kube-api-access-4dh7p\") pod \"1f72bbad-2cab-421b-9f3c-2cd27581c638\" (UID: \"1f72bbad-2cab-421b-9f3c-2cd27581c638\") " Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.271678 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f72bbad-2cab-421b-9f3c-2cd27581c638-logs" (OuterVolumeSpecName: "logs") pod "1f72bbad-2cab-421b-9f3c-2cd27581c638" (UID: "1f72bbad-2cab-421b-9f3c-2cd27581c638"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.272117 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f72bbad-2cab-421b-9f3c-2cd27581c638-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.277251 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f72bbad-2cab-421b-9f3c-2cd27581c638-kube-api-access-4dh7p" (OuterVolumeSpecName: "kube-api-access-4dh7p") pod "1f72bbad-2cab-421b-9f3c-2cd27581c638" (UID: "1f72bbad-2cab-421b-9f3c-2cd27581c638"). InnerVolumeSpecName "kube-api-access-4dh7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.298530 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f72bbad-2cab-421b-9f3c-2cd27581c638" (UID: "1f72bbad-2cab-421b-9f3c-2cd27581c638"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.309542 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-config-data" (OuterVolumeSpecName: "config-data") pod "1f72bbad-2cab-421b-9f3c-2cd27581c638" (UID: "1f72bbad-2cab-421b-9f3c-2cd27581c638"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.373881 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.373914 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f72bbad-2cab-421b-9f3c-2cd27581c638-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.373928 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dh7p\" (UniqueName: \"kubernetes.io/projected/1f72bbad-2cab-421b-9f3c-2cd27581c638-kube-api-access-4dh7p\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.543718 4744 generic.go:334] "Generic (PLEG): container finished" podID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerID="922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4" exitCode=0 Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.543750 4744 generic.go:334] "Generic (PLEG): container finished" podID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerID="75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559" exitCode=143 Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.543778 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.543792 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f72bbad-2cab-421b-9f3c-2cd27581c638","Type":"ContainerDied","Data":"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4"} Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.543888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f72bbad-2cab-421b-9f3c-2cd27581c638","Type":"ContainerDied","Data":"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559"} Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.543903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f72bbad-2cab-421b-9f3c-2cd27581c638","Type":"ContainerDied","Data":"8a6f1d14c2de2289913988be9fd35f2d3b833bae727c52c3c6b9685ad50c5eeb"} Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.543924 4744 scope.go:117] "RemoveContainer" containerID="922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.628667 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.634546 4744 scope.go:117] "RemoveContainer" containerID="75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.637557 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.650714 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:29 crc kubenswrapper[4744]: E1201 08:37:29.651155 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-log" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.651175 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-log" Dec 01 08:37:29 crc kubenswrapper[4744]: E1201 08:37:29.651192 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-metadata" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.651200 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-metadata" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.651384 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-metadata" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.651399 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" containerName="nova-metadata-log" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.652318 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.656788 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.657069 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.674071 4744 scope.go:117] "RemoveContainer" containerID="922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4" Dec 01 08:37:29 crc kubenswrapper[4744]: E1201 08:37:29.674695 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4\": container with ID starting with 922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4 not found: ID does not exist" containerID="922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.674756 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4"} err="failed to get container status \"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4\": rpc error: code = NotFound desc = could not find container \"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4\": container with ID starting with 922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4 not found: ID does not exist" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.674800 4744 scope.go:117] "RemoveContainer" containerID="75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559" Dec 01 08:37:29 crc kubenswrapper[4744]: E1201 08:37:29.675137 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559\": container with ID starting with 75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559 not found: ID does not exist" containerID="75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.675179 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559"} err="failed to get container status \"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559\": rpc error: code = NotFound desc = could not find container \"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559\": container with ID starting with 75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559 not found: ID does not exist" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.675207 4744 scope.go:117] "RemoveContainer" containerID="922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.675506 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4"} err="failed to get container status \"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4\": rpc error: code = NotFound desc = could not find container \"922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4\": container with ID starting with 922a9971ed842abf884b8c1c65b0cb692d67464864d951a3c1ed8613b6cf61d4 not found: ID does not exist" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.675543 4744 scope.go:117] "RemoveContainer" containerID="75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.677269 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559"} err="failed to get container status \"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559\": rpc error: code = NotFound desc = could not find container \"75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559\": container with ID starting with 75350224d559ce2dc75670f65e7905cbcffd34be16b45eb677d536ceec82c559 not found: ID does not exist" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.678481 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.782380 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e66581-937d-491b-ac23-8d033eca0beb-logs\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.782440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.782477 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-config-data\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.782505 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.782584 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jps5m\" (UniqueName: \"kubernetes.io/projected/c2e66581-937d-491b-ac23-8d033eca0beb-kube-api-access-jps5m\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.884899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jps5m\" (UniqueName: \"kubernetes.io/projected/c2e66581-937d-491b-ac23-8d033eca0beb-kube-api-access-jps5m\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.885052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e66581-937d-491b-ac23-8d033eca0beb-logs\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.885075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.885104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-config-data\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.885129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.886161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e66581-937d-491b-ac23-8d033eca0beb-logs\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.891372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-config-data\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.892107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.897234 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.902800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jps5m\" (UniqueName: \"kubernetes.io/projected/c2e66581-937d-491b-ac23-8d033eca0beb-kube-api-access-jps5m\") pod \"nova-metadata-0\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " pod="openstack/nova-metadata-0" Dec 01 08:37:29 crc kubenswrapper[4744]: I1201 08:37:29.972541 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:30 crc kubenswrapper[4744]: I1201 08:37:30.296053 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f72bbad-2cab-421b-9f3c-2cd27581c638" path="/var/lib/kubelet/pods/1f72bbad-2cab-421b-9f3c-2cd27581c638/volumes" Dec 01 08:37:30 crc kubenswrapper[4744]: W1201 08:37:30.444710 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2e66581_937d_491b_ac23_8d033eca0beb.slice/crio-e13216c0ba79941096fdd6dd508e95680c58147af9eb5f3f9380c98c906fb095 WatchSource:0}: Error finding container e13216c0ba79941096fdd6dd508e95680c58147af9eb5f3f9380c98c906fb095: Status 404 returned error can't find the container with id e13216c0ba79941096fdd6dd508e95680c58147af9eb5f3f9380c98c906fb095 Dec 01 08:37:30 crc kubenswrapper[4744]: I1201 08:37:30.444749 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:30 crc kubenswrapper[4744]: I1201 08:37:30.562702 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e66581-937d-491b-ac23-8d033eca0beb","Type":"ContainerStarted","Data":"e13216c0ba79941096fdd6dd508e95680c58147af9eb5f3f9380c98c906fb095"} Dec 01 08:37:31 crc kubenswrapper[4744]: I1201 08:37:31.580366 4744 generic.go:334] "Generic (PLEG): container finished" podID="040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" containerID="789cde6612119cbc9df1a36a57e6970528223e416fbfee18e1cb1cfe62d93173" exitCode=0 Dec 01 08:37:31 crc kubenswrapper[4744]: I1201 08:37:31.580450 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jc998" event={"ID":"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf","Type":"ContainerDied","Data":"789cde6612119cbc9df1a36a57e6970528223e416fbfee18e1cb1cfe62d93173"} Dec 01 08:37:31 crc kubenswrapper[4744]: I1201 08:37:31.583981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e66581-937d-491b-ac23-8d033eca0beb","Type":"ContainerStarted","Data":"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05"} Dec 01 08:37:31 crc kubenswrapper[4744]: I1201 08:37:31.584646 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e66581-937d-491b-ac23-8d033eca0beb","Type":"ContainerStarted","Data":"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c"} Dec 01 08:37:31 crc kubenswrapper[4744]: I1201 08:37:31.587552 4744 generic.go:334] "Generic (PLEG): container finished" podID="3eb87b97-a941-413c-9778-29a593d44ef0" containerID="b2a2a496c6333eb838de95bcb18cdb525d89b52ecaaeee3af278068c84bf73bf" exitCode=0 Dec 01 08:37:31 crc kubenswrapper[4744]: I1201 08:37:31.587600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-24cgv" event={"ID":"3eb87b97-a941-413c-9778-29a593d44ef0","Type":"ContainerDied","Data":"b2a2a496c6333eb838de95bcb18cdb525d89b52ecaaeee3af278068c84bf73bf"} Dec 01 08:37:31 crc kubenswrapper[4744]: I1201 08:37:31.666003 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.665972118 podStartE2EDuration="2.665972118s" podCreationTimestamp="2025-12-01 08:37:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:31.659575056 +0000 UTC m=+1183.648633007" watchObservedRunningTime="2025-12-01 08:37:31.665972118 +0000 UTC m=+1183.655030069" Dec 01 08:37:32 crc kubenswrapper[4744]: I1201 08:37:32.916660 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 08:37:32 crc kubenswrapper[4744]: I1201 08:37:32.916942 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.059016 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.059251 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.071944 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.087277 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.122642 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.143606 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-zqdmr"] Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.143789 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" podUID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerName="dnsmasq-dns" containerID="cri-o://ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8" gracePeriod=10 Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.145241 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.158787 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.269627 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-scripts\") pod \"3eb87b97-a941-413c-9778-29a593d44ef0\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.269689 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-scripts\") pod \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.269764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-config-data\") pod \"3eb87b97-a941-413c-9778-29a593d44ef0\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.269808 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-combined-ca-bundle\") pod \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.269876 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-config-data\") pod \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.269923 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpfcv\" (UniqueName: \"kubernetes.io/projected/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-kube-api-access-mpfcv\") pod \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\" (UID: \"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.269970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh4kl\" (UniqueName: \"kubernetes.io/projected/3eb87b97-a941-413c-9778-29a593d44ef0-kube-api-access-jh4kl\") pod \"3eb87b97-a941-413c-9778-29a593d44ef0\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.270019 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-combined-ca-bundle\") pod \"3eb87b97-a941-413c-9778-29a593d44ef0\" (UID: \"3eb87b97-a941-413c-9778-29a593d44ef0\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.280718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-kube-api-access-mpfcv" (OuterVolumeSpecName: "kube-api-access-mpfcv") pod "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" (UID: "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf"). InnerVolumeSpecName "kube-api-access-mpfcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.283532 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-scripts" (OuterVolumeSpecName: "scripts") pod "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" (UID: "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.283591 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-scripts" (OuterVolumeSpecName: "scripts") pod "3eb87b97-a941-413c-9778-29a593d44ef0" (UID: "3eb87b97-a941-413c-9778-29a593d44ef0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.297684 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eb87b97-a941-413c-9778-29a593d44ef0-kube-api-access-jh4kl" (OuterVolumeSpecName: "kube-api-access-jh4kl") pod "3eb87b97-a941-413c-9778-29a593d44ef0" (UID: "3eb87b97-a941-413c-9778-29a593d44ef0"). InnerVolumeSpecName "kube-api-access-jh4kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.318117 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3eb87b97-a941-413c-9778-29a593d44ef0" (UID: "3eb87b97-a941-413c-9778-29a593d44ef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.327825 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-config-data" (OuterVolumeSpecName: "config-data") pod "3eb87b97-a941-413c-9778-29a593d44ef0" (UID: "3eb87b97-a941-413c-9778-29a593d44ef0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.347548 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-config-data" (OuterVolumeSpecName: "config-data") pod "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" (UID: "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.348688 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" (UID: "040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373600 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh4kl\" (UniqueName: \"kubernetes.io/projected/3eb87b97-a941-413c-9778-29a593d44ef0-kube-api-access-jh4kl\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373638 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373647 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373656 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373664 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eb87b97-a941-413c-9778-29a593d44ef0-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373673 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373682 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.373690 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpfcv\" (UniqueName: \"kubernetes.io/projected/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf-kube-api-access-mpfcv\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.551342 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.609259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-24cgv" event={"ID":"3eb87b97-a941-413c-9778-29a593d44ef0","Type":"ContainerDied","Data":"d900dbc4a27a04d16bafade8317a9e12d856187b9d3ee0f6390194e5433d9cb1"} Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.609297 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d900dbc4a27a04d16bafade8317a9e12d856187b9d3ee0f6390194e5433d9cb1" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.609352 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-24cgv" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.621589 4744 generic.go:334] "Generic (PLEG): container finished" podID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerID="ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8" exitCode=0 Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.621696 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.621699 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" event={"ID":"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb","Type":"ContainerDied","Data":"ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8"} Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.621814 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-zqdmr" event={"ID":"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb","Type":"ContainerDied","Data":"4996bc9300e8487e347311584a654f937b8cbada6f45e708d70b201ba7aede5a"} Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.621835 4744 scope.go:117] "RemoveContainer" containerID="ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.626466 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jc998" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.626505 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jc998" event={"ID":"040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf","Type":"ContainerDied","Data":"0b0bff0e0f14a5233cc609c038fdef1be5df5298cbb614cde9679155ed73b37a"} Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.626543 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b0bff0e0f14a5233cc609c038fdef1be5df5298cbb614cde9679155ed73b37a" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.671434 4744 scope.go:117] "RemoveContainer" containerID="cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.682373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-svc\") pod \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.682448 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-sb\") pod \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.682553 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-config\") pod \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.682602 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp5pt\" (UniqueName: \"kubernetes.io/projected/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-kube-api-access-sp5pt\") pod \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.682732 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-nb\") pod \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.682801 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-swift-storage-0\") pod \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\" (UID: \"aae079b7-b2c8-49f5-80e1-c3e1e04c7edb\") " Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.687331 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-kube-api-access-sp5pt" (OuterVolumeSpecName: "kube-api-access-sp5pt") pod "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" (UID: "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb"). InnerVolumeSpecName "kube-api-access-sp5pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.687555 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.699708 4744 scope.go:117] "RemoveContainer" containerID="ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8" Dec 01 08:37:33 crc kubenswrapper[4744]: E1201 08:37:33.708094 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8\": container with ID starting with ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8 not found: ID does not exist" containerID="ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.708140 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8"} err="failed to get container status \"ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8\": rpc error: code = NotFound desc = could not find container \"ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8\": container with ID starting with ef817ff9b9cc3c09868921f8d9d6aa2fc8842b220ddf94b606bb7ac9df2454d8 not found: ID does not exist" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.708167 4744 scope.go:117] "RemoveContainer" containerID="cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac" Dec 01 08:37:33 crc kubenswrapper[4744]: E1201 08:37:33.708438 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac\": container with ID starting with cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac not found: ID does not exist" containerID="cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.708456 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac"} err="failed to get container status \"cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac\": rpc error: code = NotFound desc = could not find container \"cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac\": container with ID starting with cb6d93530df312c30cd71e5567bea5d3852523b2ade4989bcbe5a8c47ab479ac not found: ID does not exist" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.718465 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 08:37:33 crc kubenswrapper[4744]: E1201 08:37:33.718900 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb87b97-a941-413c-9778-29a593d44ef0" containerName="nova-cell1-conductor-db-sync" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.718917 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb87b97-a941-413c-9778-29a593d44ef0" containerName="nova-cell1-conductor-db-sync" Dec 01 08:37:33 crc kubenswrapper[4744]: E1201 08:37:33.718931 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerName="dnsmasq-dns" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.718938 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerName="dnsmasq-dns" Dec 01 08:37:33 crc kubenswrapper[4744]: E1201 08:37:33.718948 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerName="init" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.718954 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerName="init" Dec 01 08:37:33 crc kubenswrapper[4744]: E1201 08:37:33.718966 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" containerName="nova-manage" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.718972 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" containerName="nova-manage" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.719146 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" containerName="dnsmasq-dns" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.719165 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eb87b97-a941-413c-9778-29a593d44ef0" containerName="nova-cell1-conductor-db-sync" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.719181 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" containerName="nova-manage" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.719885 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.724929 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.737919 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.754718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" (UID: "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.764719 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-config" (OuterVolumeSpecName: "config") pod "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" (UID: "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.768285 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" (UID: "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.771921 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" (UID: "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.772141 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" (UID: "aae079b7-b2c8-49f5-80e1-c3e1e04c7edb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.785306 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.785341 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.785354 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.785364 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.785374 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp5pt\" (UniqueName: \"kubernetes.io/projected/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-kube-api-access-sp5pt\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.785384 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.823321 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.823690 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-api" containerID="cri-o://853334c399072bc499caa4ab92ef6484754bcfcdf9ad564e62a5b3663f302364" gracePeriod=30 Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.823564 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-log" containerID="cri-o://2557df3e2a35ca8c183f221af58573cea9a46f7dbcb7ebcf551e0a308534a886" gracePeriod=30 Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.827506 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": EOF" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.827862 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": EOF" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.866163 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.866364 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-log" containerID="cri-o://45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c" gracePeriod=30 Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.866765 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-metadata" containerID="cri-o://a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05" gracePeriod=30 Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.887269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.887337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8zr8\" (UniqueName: \"kubernetes.io/projected/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-kube-api-access-w8zr8\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.887360 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.953203 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-zqdmr"] Dec 01 08:37:33 crc kubenswrapper[4744]: I1201 08:37:33.960510 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-zqdmr"] Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.009589 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.009994 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8zr8\" (UniqueName: \"kubernetes.io/projected/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-kube-api-access-w8zr8\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.010083 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.018151 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.022894 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.042131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8zr8\" (UniqueName: \"kubernetes.io/projected/0ae053c1-f41c-4bbb-bae8-66e1d3d891b7-kube-api-access-w8zr8\") pod \"nova-cell1-conductor-0\" (UID: \"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7\") " pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.046558 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.300132 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae079b7-b2c8-49f5-80e1-c3e1e04c7edb" path="/var/lib/kubelet/pods/aae079b7-b2c8-49f5-80e1-c3e1e04c7edb/volumes" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.309599 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.575492 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.575772 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 08:37:34 crc kubenswrapper[4744]: W1201 08:37:34.581939 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ae053c1_f41c_4bbb_bae8_66e1d3d891b7.slice/crio-862e74cb8916120182f3b4ef69dffa81a8fbbdbb24d63d1fc1a7f60c2e9a3937 WatchSource:0}: Error finding container 862e74cb8916120182f3b4ef69dffa81a8fbbdbb24d63d1fc1a7f60c2e9a3937: Status 404 returned error can't find the container with id 862e74cb8916120182f3b4ef69dffa81a8fbbdbb24d63d1fc1a7f60c2e9a3937 Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.636480 4744 generic.go:334] "Generic (PLEG): container finished" podID="c2e66581-937d-491b-ac23-8d033eca0beb" containerID="a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05" exitCode=0 Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.636763 4744 generic.go:334] "Generic (PLEG): container finished" podID="c2e66581-937d-491b-ac23-8d033eca0beb" containerID="45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c" exitCode=143 Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.636838 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e66581-937d-491b-ac23-8d033eca0beb","Type":"ContainerDied","Data":"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05"} Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.636864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e66581-937d-491b-ac23-8d033eca0beb","Type":"ContainerDied","Data":"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c"} Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.636874 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2e66581-937d-491b-ac23-8d033eca0beb","Type":"ContainerDied","Data":"e13216c0ba79941096fdd6dd508e95680c58147af9eb5f3f9380c98c906fb095"} Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.636890 4744 scope.go:117] "RemoveContainer" containerID="a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.637305 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.642515 4744 generic.go:334] "Generic (PLEG): container finished" podID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerID="2557df3e2a35ca8c183f221af58573cea9a46f7dbcb7ebcf551e0a308534a886" exitCode=143 Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.642571 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce3cbc79-3564-42d0-9005-a9c503c25f00","Type":"ContainerDied","Data":"2557df3e2a35ca8c183f221af58573cea9a46f7dbcb7ebcf551e0a308534a886"} Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.643598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7","Type":"ContainerStarted","Data":"862e74cb8916120182f3b4ef69dffa81a8fbbdbb24d63d1fc1a7f60c2e9a3937"} Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.667865 4744 scope.go:117] "RemoveContainer" containerID="45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.688932 4744 scope.go:117] "RemoveContainer" containerID="a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05" Dec 01 08:37:34 crc kubenswrapper[4744]: E1201 08:37:34.692017 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05\": container with ID starting with a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05 not found: ID does not exist" containerID="a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.692095 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05"} err="failed to get container status \"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05\": rpc error: code = NotFound desc = could not find container \"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05\": container with ID starting with a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05 not found: ID does not exist" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.692135 4744 scope.go:117] "RemoveContainer" containerID="45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c" Dec 01 08:37:34 crc kubenswrapper[4744]: E1201 08:37:34.692790 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c\": container with ID starting with 45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c not found: ID does not exist" containerID="45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.692830 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c"} err="failed to get container status \"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c\": rpc error: code = NotFound desc = could not find container \"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c\": container with ID starting with 45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c not found: ID does not exist" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.692849 4744 scope.go:117] "RemoveContainer" containerID="a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.694028 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05"} err="failed to get container status \"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05\": rpc error: code = NotFound desc = could not find container \"a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05\": container with ID starting with a9fa1060bbb0b8290dabd7af0d1bcad7b04b1d6a16c5b740f543f0060f5f9e05 not found: ID does not exist" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.694056 4744 scope.go:117] "RemoveContainer" containerID="45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.694495 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c"} err="failed to get container status \"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c\": rpc error: code = NotFound desc = could not find container \"45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c\": container with ID starting with 45bda6f6431ddba1f58b01512ee27945ef1487984ddb78021ec69f254c16c42c not found: ID does not exist" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.721349 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e66581-937d-491b-ac23-8d033eca0beb-logs\") pod \"c2e66581-937d-491b-ac23-8d033eca0beb\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.721461 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jps5m\" (UniqueName: \"kubernetes.io/projected/c2e66581-937d-491b-ac23-8d033eca0beb-kube-api-access-jps5m\") pod \"c2e66581-937d-491b-ac23-8d033eca0beb\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.721549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-combined-ca-bundle\") pod \"c2e66581-937d-491b-ac23-8d033eca0beb\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.721599 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-config-data\") pod \"c2e66581-937d-491b-ac23-8d033eca0beb\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.721679 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-nova-metadata-tls-certs\") pod \"c2e66581-937d-491b-ac23-8d033eca0beb\" (UID: \"c2e66581-937d-491b-ac23-8d033eca0beb\") " Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.724049 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2e66581-937d-491b-ac23-8d033eca0beb-logs" (OuterVolumeSpecName: "logs") pod "c2e66581-937d-491b-ac23-8d033eca0beb" (UID: "c2e66581-937d-491b-ac23-8d033eca0beb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.735144 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e66581-937d-491b-ac23-8d033eca0beb-kube-api-access-jps5m" (OuterVolumeSpecName: "kube-api-access-jps5m") pod "c2e66581-937d-491b-ac23-8d033eca0beb" (UID: "c2e66581-937d-491b-ac23-8d033eca0beb"). InnerVolumeSpecName "kube-api-access-jps5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.762377 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2e66581-937d-491b-ac23-8d033eca0beb" (UID: "c2e66581-937d-491b-ac23-8d033eca0beb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.765309 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-config-data" (OuterVolumeSpecName: "config-data") pod "c2e66581-937d-491b-ac23-8d033eca0beb" (UID: "c2e66581-937d-491b-ac23-8d033eca0beb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.811297 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c2e66581-937d-491b-ac23-8d033eca0beb" (UID: "c2e66581-937d-491b-ac23-8d033eca0beb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.824353 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2e66581-937d-491b-ac23-8d033eca0beb-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.824598 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jps5m\" (UniqueName: \"kubernetes.io/projected/c2e66581-937d-491b-ac23-8d033eca0beb-kube-api-access-jps5m\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.824660 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.824676 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.824689 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e66581-937d-491b-ac23-8d033eca0beb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.976571 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:34 crc kubenswrapper[4744]: I1201 08:37:34.996054 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.043039 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:35 crc kubenswrapper[4744]: E1201 08:37:35.043434 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-log" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.043451 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-log" Dec 01 08:37:35 crc kubenswrapper[4744]: E1201 08:37:35.043488 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-metadata" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.043495 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-metadata" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.043652 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-log" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.043666 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" containerName="nova-metadata-metadata" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.044565 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.047069 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.049201 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.066353 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.130119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-config-data\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.130219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.130363 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prv6c\" (UniqueName: \"kubernetes.io/projected/19dbb279-f666-4a27-a699-88ccd0b5cf2b-kube-api-access-prv6c\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.130395 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19dbb279-f666-4a27-a699-88ccd0b5cf2b-logs\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.130535 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.233635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-config-data\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.234261 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.234463 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prv6c\" (UniqueName: \"kubernetes.io/projected/19dbb279-f666-4a27-a699-88ccd0b5cf2b-kube-api-access-prv6c\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.234583 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19dbb279-f666-4a27-a699-88ccd0b5cf2b-logs\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.234763 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.235195 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19dbb279-f666-4a27-a699-88ccd0b5cf2b-logs\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.246205 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-config-data\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.252641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prv6c\" (UniqueName: \"kubernetes.io/projected/19dbb279-f666-4a27-a699-88ccd0b5cf2b-kube-api-access-prv6c\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.252845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.254304 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.373292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.654611 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0ae053c1-f41c-4bbb-bae8-66e1d3d891b7","Type":"ContainerStarted","Data":"fde4033e4aad72637f0bdfa0c3d09bf0594772c50e4bb88a790cac2ec2237805"} Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.654892 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.654768 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="92d02f35-1811-4824-a1c8-375ebde20065" containerName="nova-scheduler-scheduler" containerID="cri-o://a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e" gracePeriod=30 Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.679568 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.679550088 podStartE2EDuration="2.679550088s" podCreationTimestamp="2025-12-01 08:37:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:35.669062898 +0000 UTC m=+1187.658120829" watchObservedRunningTime="2025-12-01 08:37:35.679550088 +0000 UTC m=+1187.668608009" Dec 01 08:37:35 crc kubenswrapper[4744]: I1201 08:37:35.835486 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:37:35 crc kubenswrapper[4744]: W1201 08:37:35.855829 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19dbb279_f666_4a27_a699_88ccd0b5cf2b.slice/crio-8425bb9558f15b8cb90598cca9e4767f9d6bc701a8623b15119b6bce88cc77c1 WatchSource:0}: Error finding container 8425bb9558f15b8cb90598cca9e4767f9d6bc701a8623b15119b6bce88cc77c1: Status 404 returned error can't find the container with id 8425bb9558f15b8cb90598cca9e4767f9d6bc701a8623b15119b6bce88cc77c1 Dec 01 08:37:36 crc kubenswrapper[4744]: I1201 08:37:36.296642 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2e66581-937d-491b-ac23-8d033eca0beb" path="/var/lib/kubelet/pods/c2e66581-937d-491b-ac23-8d033eca0beb/volumes" Dec 01 08:37:36 crc kubenswrapper[4744]: I1201 08:37:36.667369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19dbb279-f666-4a27-a699-88ccd0b5cf2b","Type":"ContainerStarted","Data":"9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd"} Dec 01 08:37:36 crc kubenswrapper[4744]: I1201 08:37:36.669500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19dbb279-f666-4a27-a699-88ccd0b5cf2b","Type":"ContainerStarted","Data":"b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059"} Dec 01 08:37:36 crc kubenswrapper[4744]: I1201 08:37:36.669695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19dbb279-f666-4a27-a699-88ccd0b5cf2b","Type":"ContainerStarted","Data":"8425bb9558f15b8cb90598cca9e4767f9d6bc701a8623b15119b6bce88cc77c1"} Dec 01 08:37:36 crc kubenswrapper[4744]: I1201 08:37:36.695629 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.695604185 podStartE2EDuration="1.695604185s" podCreationTimestamp="2025-12-01 08:37:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:36.684305963 +0000 UTC m=+1188.673363874" watchObservedRunningTime="2025-12-01 08:37:36.695604185 +0000 UTC m=+1188.684662106" Dec 01 08:37:38 crc kubenswrapper[4744]: E1201 08:37:38.061458 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 08:37:38 crc kubenswrapper[4744]: E1201 08:37:38.064185 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 08:37:38 crc kubenswrapper[4744]: E1201 08:37:38.067316 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 08:37:38 crc kubenswrapper[4744]: E1201 08:37:38.067457 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="92d02f35-1811-4824-a1c8-375ebde20065" containerName="nova-scheduler-scheduler" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.095325 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.699763 4744 generic.go:334] "Generic (PLEG): container finished" podID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerID="853334c399072bc499caa4ab92ef6484754bcfcdf9ad564e62a5b3663f302364" exitCode=0 Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.699837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce3cbc79-3564-42d0-9005-a9c503c25f00","Type":"ContainerDied","Data":"853334c399072bc499caa4ab92ef6484754bcfcdf9ad564e62a5b3663f302364"} Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.700073 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce3cbc79-3564-42d0-9005-a9c503c25f00","Type":"ContainerDied","Data":"68831277585c33f25a037fa64b6a32238dfc9c21c04388a2383d6fd69d87e604"} Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.700091 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68831277585c33f25a037fa64b6a32238dfc9c21c04388a2383d6fd69d87e604" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.701491 4744 generic.go:334] "Generic (PLEG): container finished" podID="92d02f35-1811-4824-a1c8-375ebde20065" containerID="a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e" exitCode=0 Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.701524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92d02f35-1811-4824-a1c8-375ebde20065","Type":"ContainerDied","Data":"a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e"} Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.770819 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.933486 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.934343 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3cbc79-3564-42d0-9005-a9c503c25f00-logs\") pod \"ce3cbc79-3564-42d0-9005-a9c503c25f00\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.934425 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-combined-ca-bundle\") pod \"ce3cbc79-3564-42d0-9005-a9c503c25f00\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.934484 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-config-data\") pod \"ce3cbc79-3564-42d0-9005-a9c503c25f00\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.934567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtnvt\" (UniqueName: \"kubernetes.io/projected/ce3cbc79-3564-42d0-9005-a9c503c25f00-kube-api-access-rtnvt\") pod \"ce3cbc79-3564-42d0-9005-a9c503c25f00\" (UID: \"ce3cbc79-3564-42d0-9005-a9c503c25f00\") " Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.934817 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce3cbc79-3564-42d0-9005-a9c503c25f00-logs" (OuterVolumeSpecName: "logs") pod "ce3cbc79-3564-42d0-9005-a9c503c25f00" (UID: "ce3cbc79-3564-42d0-9005-a9c503c25f00"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.935213 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3cbc79-3564-42d0-9005-a9c503c25f00-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.941669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3cbc79-3564-42d0-9005-a9c503c25f00-kube-api-access-rtnvt" (OuterVolumeSpecName: "kube-api-access-rtnvt") pod "ce3cbc79-3564-42d0-9005-a9c503c25f00" (UID: "ce3cbc79-3564-42d0-9005-a9c503c25f00"). InnerVolumeSpecName "kube-api-access-rtnvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.965739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-config-data" (OuterVolumeSpecName: "config-data") pod "ce3cbc79-3564-42d0-9005-a9c503c25f00" (UID: "ce3cbc79-3564-42d0-9005-a9c503c25f00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:39 crc kubenswrapper[4744]: I1201 08:37:39.979463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce3cbc79-3564-42d0-9005-a9c503c25f00" (UID: "ce3cbc79-3564-42d0-9005-a9c503c25f00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.036642 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-combined-ca-bundle\") pod \"92d02f35-1811-4824-a1c8-375ebde20065\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.036717 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-config-data\") pod \"92d02f35-1811-4824-a1c8-375ebde20065\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.036971 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w59bl\" (UniqueName: \"kubernetes.io/projected/92d02f35-1811-4824-a1c8-375ebde20065-kube-api-access-w59bl\") pod \"92d02f35-1811-4824-a1c8-375ebde20065\" (UID: \"92d02f35-1811-4824-a1c8-375ebde20065\") " Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.037503 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.037525 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtnvt\" (UniqueName: \"kubernetes.io/projected/ce3cbc79-3564-42d0-9005-a9c503c25f00-kube-api-access-rtnvt\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.037538 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3cbc79-3564-42d0-9005-a9c503c25f00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.041118 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d02f35-1811-4824-a1c8-375ebde20065-kube-api-access-w59bl" (OuterVolumeSpecName: "kube-api-access-w59bl") pod "92d02f35-1811-4824-a1c8-375ebde20065" (UID: "92d02f35-1811-4824-a1c8-375ebde20065"). InnerVolumeSpecName "kube-api-access-w59bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.067774 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-config-data" (OuterVolumeSpecName: "config-data") pod "92d02f35-1811-4824-a1c8-375ebde20065" (UID: "92d02f35-1811-4824-a1c8-375ebde20065"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.069341 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92d02f35-1811-4824-a1c8-375ebde20065" (UID: "92d02f35-1811-4824-a1c8-375ebde20065"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.139015 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w59bl\" (UniqueName: \"kubernetes.io/projected/92d02f35-1811-4824-a1c8-375ebde20065-kube-api-access-w59bl\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.139061 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.139072 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d02f35-1811-4824-a1c8-375ebde20065-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.374018 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.375485 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.713553 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92d02f35-1811-4824-a1c8-375ebde20065","Type":"ContainerDied","Data":"8330e6a909b874935025c0442c9ff5a31614484702eaf1eba6a3c86b26be050d"} Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.713655 4744 scope.go:117] "RemoveContainer" containerID="a0926ea99f49bbaf46b50ee8c89befdf89388833f3e6a028bb75448d1dcda90e" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.713688 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.713592 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.757368 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.773632 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.789254 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.803560 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.819493 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: E1201 08:37:40.819920 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d02f35-1811-4824-a1c8-375ebde20065" containerName="nova-scheduler-scheduler" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.819942 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d02f35-1811-4824-a1c8-375ebde20065" containerName="nova-scheduler-scheduler" Dec 01 08:37:40 crc kubenswrapper[4744]: E1201 08:37:40.819992 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-log" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.820001 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-log" Dec 01 08:37:40 crc kubenswrapper[4744]: E1201 08:37:40.820013 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-api" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.820022 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-api" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.820231 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d02f35-1811-4824-a1c8-375ebde20065" containerName="nova-scheduler-scheduler" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.820265 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-log" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.820283 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" containerName="nova-api-api" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.824033 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.826560 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.838921 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.840269 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.842510 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.853690 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.864954 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.980179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmbq9\" (UniqueName: \"kubernetes.io/projected/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-kube-api-access-pmbq9\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.980387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.980520 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-config-data\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.980586 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.980797 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74fdf4b-20a8-4df0-96c0-ff5e06339169-logs\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.980985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-config-data\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:40 crc kubenswrapper[4744]: I1201 08:37:40.981048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfh24\" (UniqueName: \"kubernetes.io/projected/d74fdf4b-20a8-4df0-96c0-ff5e06339169-kube-api-access-tfh24\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.083017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmbq9\" (UniqueName: \"kubernetes.io/projected/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-kube-api-access-pmbq9\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.083285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.083375 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-config-data\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.083479 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.083579 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74fdf4b-20a8-4df0-96c0-ff5e06339169-logs\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.083692 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-config-data\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.083773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfh24\" (UniqueName: \"kubernetes.io/projected/d74fdf4b-20a8-4df0-96c0-ff5e06339169-kube-api-access-tfh24\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.084053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74fdf4b-20a8-4df0-96c0-ff5e06339169-logs\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.092778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.093357 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-config-data\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.093895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.094161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-config-data\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.098729 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmbq9\" (UniqueName: \"kubernetes.io/projected/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-kube-api-access-pmbq9\") pod \"nova-scheduler-0\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " pod="openstack/nova-scheduler-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.109427 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfh24\" (UniqueName: \"kubernetes.io/projected/d74fdf4b-20a8-4df0-96c0-ff5e06339169-kube-api-access-tfh24\") pod \"nova-api-0\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.143663 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.156924 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.650731 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.661024 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:37:41 crc kubenswrapper[4744]: W1201 08:37:41.663835 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bb988e1_ef1a_45b8_b3fc_cf6025835a89.slice/crio-94c07d76039e5dd41953446ecbeceb3f7a3f6d3b5fb95da5d1ce0e80d166f766 WatchSource:0}: Error finding container 94c07d76039e5dd41953446ecbeceb3f7a3f6d3b5fb95da5d1ce0e80d166f766: Status 404 returned error can't find the container with id 94c07d76039e5dd41953446ecbeceb3f7a3f6d3b5fb95da5d1ce0e80d166f766 Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.725866 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bb988e1-ef1a-45b8-b3fc-cf6025835a89","Type":"ContainerStarted","Data":"94c07d76039e5dd41953446ecbeceb3f7a3f6d3b5fb95da5d1ce0e80d166f766"} Dec 01 08:37:41 crc kubenswrapper[4744]: I1201 08:37:41.728824 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d74fdf4b-20a8-4df0-96c0-ff5e06339169","Type":"ContainerStarted","Data":"997577237a5d66040871dfb92561ab5651a401e12cb7ace550d091f55bbfeee8"} Dec 01 08:37:42 crc kubenswrapper[4744]: I1201 08:37:42.296771 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d02f35-1811-4824-a1c8-375ebde20065" path="/var/lib/kubelet/pods/92d02f35-1811-4824-a1c8-375ebde20065/volumes" Dec 01 08:37:42 crc kubenswrapper[4744]: I1201 08:37:42.297332 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3cbc79-3564-42d0-9005-a9c503c25f00" path="/var/lib/kubelet/pods/ce3cbc79-3564-42d0-9005-a9c503c25f00/volumes" Dec 01 08:37:42 crc kubenswrapper[4744]: I1201 08:37:42.741467 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d74fdf4b-20a8-4df0-96c0-ff5e06339169","Type":"ContainerStarted","Data":"fc68715a13f1dc8c8706e40aa2f9d1f7c76b852168d83a09ac1025d5aed208a9"} Dec 01 08:37:42 crc kubenswrapper[4744]: I1201 08:37:42.741866 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d74fdf4b-20a8-4df0-96c0-ff5e06339169","Type":"ContainerStarted","Data":"b5629edf195a5d2c2e573937fb9b00785104f6e528f26e46fd3b40e0aed429a4"} Dec 01 08:37:42 crc kubenswrapper[4744]: I1201 08:37:42.744159 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bb988e1-ef1a-45b8-b3fc-cf6025835a89","Type":"ContainerStarted","Data":"f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb"} Dec 01 08:37:42 crc kubenswrapper[4744]: I1201 08:37:42.766648 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.766624855 podStartE2EDuration="2.766624855s" podCreationTimestamp="2025-12-01 08:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:42.761362424 +0000 UTC m=+1194.750420385" watchObservedRunningTime="2025-12-01 08:37:42.766624855 +0000 UTC m=+1194.755682786" Dec 01 08:37:42 crc kubenswrapper[4744]: I1201 08:37:42.807086 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.807062128 podStartE2EDuration="2.807062128s" podCreationTimestamp="2025-12-01 08:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:37:42.787446669 +0000 UTC m=+1194.776504600" watchObservedRunningTime="2025-12-01 08:37:42.807062128 +0000 UTC m=+1194.796120059" Dec 01 08:37:45 crc kubenswrapper[4744]: I1201 08:37:45.374583 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 08:37:45 crc kubenswrapper[4744]: I1201 08:37:45.375093 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 08:37:46 crc kubenswrapper[4744]: I1201 08:37:46.157399 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 08:37:46 crc kubenswrapper[4744]: I1201 08:37:46.391667 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 08:37:46 crc kubenswrapper[4744]: I1201 08:37:46.391673 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 08:37:47 crc kubenswrapper[4744]: I1201 08:37:47.040039 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 08:37:48 crc kubenswrapper[4744]: I1201 08:37:48.635450 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:37:48 crc kubenswrapper[4744]: I1201 08:37:48.635812 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:37:50 crc kubenswrapper[4744]: I1201 08:37:50.789721 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:37:50 crc kubenswrapper[4744]: I1201 08:37:50.790298 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e70a50a9-9bb9-45be-aa60-c57c67e9fc32" containerName="kube-state-metrics" containerID="cri-o://089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680" gracePeriod=30 Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.144937 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.145256 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.157599 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.188671 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.273945 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.432177 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4mrh\" (UniqueName: \"kubernetes.io/projected/e70a50a9-9bb9-45be-aa60-c57c67e9fc32-kube-api-access-z4mrh\") pod \"e70a50a9-9bb9-45be-aa60-c57c67e9fc32\" (UID: \"e70a50a9-9bb9-45be-aa60-c57c67e9fc32\") " Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.437747 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70a50a9-9bb9-45be-aa60-c57c67e9fc32-kube-api-access-z4mrh" (OuterVolumeSpecName: "kube-api-access-z4mrh") pod "e70a50a9-9bb9-45be-aa60-c57c67e9fc32" (UID: "e70a50a9-9bb9-45be-aa60-c57c67e9fc32"). InnerVolumeSpecName "kube-api-access-z4mrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.535902 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4mrh\" (UniqueName: \"kubernetes.io/projected/e70a50a9-9bb9-45be-aa60-c57c67e9fc32-kube-api-access-z4mrh\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.858462 4744 generic.go:334] "Generic (PLEG): container finished" podID="e70a50a9-9bb9-45be-aa60-c57c67e9fc32" containerID="089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680" exitCode=2 Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.859512 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.862771 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e70a50a9-9bb9-45be-aa60-c57c67e9fc32","Type":"ContainerDied","Data":"089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680"} Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.862843 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e70a50a9-9bb9-45be-aa60-c57c67e9fc32","Type":"ContainerDied","Data":"26ebc04757163c14cb46ef58235eeb675be3c384777e27f94144c670346f7d9c"} Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.862894 4744 scope.go:117] "RemoveContainer" containerID="089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.906698 4744 scope.go:117] "RemoveContainer" containerID="089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680" Dec 01 08:37:51 crc kubenswrapper[4744]: E1201 08:37:51.907449 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680\": container with ID starting with 089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680 not found: ID does not exist" containerID="089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.907510 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680"} err="failed to get container status \"089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680\": rpc error: code = NotFound desc = could not find container \"089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680\": container with ID starting with 089210cfb479d20e23c48d740991f734d75f9de4679ca864750c4c9d2f0af680 not found: ID does not exist" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.912291 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.930471 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.946915 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.961258 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:37:51 crc kubenswrapper[4744]: E1201 08:37:51.961763 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70a50a9-9bb9-45be-aa60-c57c67e9fc32" containerName="kube-state-metrics" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.961788 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70a50a9-9bb9-45be-aa60-c57c67e9fc32" containerName="kube-state-metrics" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.962031 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70a50a9-9bb9-45be-aa60-c57c67e9fc32" containerName="kube-state-metrics" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.962829 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.968619 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.968673 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 08:37:51 crc kubenswrapper[4744]: I1201 08:37:51.981806 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.149582 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zldzm\" (UniqueName: \"kubernetes.io/projected/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-api-access-zldzm\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.149645 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.149682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.149721 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.186886 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.186855 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.251245 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.251346 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.251470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zldzm\" (UniqueName: \"kubernetes.io/projected/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-api-access-zldzm\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.251517 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.257884 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.258755 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.263860 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.289623 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zldzm\" (UniqueName: \"kubernetes.io/projected/13bbf1b1-4680-4390-99ed-e0a54dd6d6fe-kube-api-access-zldzm\") pod \"kube-state-metrics-0\" (UID: \"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe\") " pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.293395 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.313092 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70a50a9-9bb9-45be-aa60-c57c67e9fc32" path="/var/lib/kubelet/pods/e70a50a9-9bb9-45be-aa60-c57c67e9fc32/volumes" Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.722509 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.722850 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-central-agent" containerID="cri-o://53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0" gracePeriod=30 Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.722913 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="sg-core" containerID="cri-o://2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b" gracePeriod=30 Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.722937 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-notification-agent" containerID="cri-o://1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d" gracePeriod=30 Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.723135 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="proxy-httpd" containerID="cri-o://c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae" gracePeriod=30 Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.804541 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 08:37:52 crc kubenswrapper[4744]: I1201 08:37:52.910685 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe","Type":"ContainerStarted","Data":"9c1d39a375b3a736ab918081dbf0a4295afc9d4baa2b5a870573388d3de4ad82"} Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.921070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"13bbf1b1-4680-4390-99ed-e0a54dd6d6fe","Type":"ContainerStarted","Data":"a19f451f466691a63d10fe22fdcbad9d0ed6043bcc9c29aed8af79e5d9132bd8"} Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.921576 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.923821 4744 generic.go:334] "Generic (PLEG): container finished" podID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerID="c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae" exitCode=0 Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.923853 4744 generic.go:334] "Generic (PLEG): container finished" podID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerID="2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b" exitCode=2 Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.923864 4744 generic.go:334] "Generic (PLEG): container finished" podID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerID="53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0" exitCode=0 Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.923876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerDied","Data":"c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae"} Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.923903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerDied","Data":"2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b"} Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.923914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerDied","Data":"53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0"} Dec 01 08:37:53 crc kubenswrapper[4744]: I1201 08:37:53.942782 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.463407798 podStartE2EDuration="2.942764544s" podCreationTimestamp="2025-12-01 08:37:51 +0000 UTC" firstStartedPulling="2025-12-01 08:37:52.81256276 +0000 UTC m=+1204.801620681" lastFinishedPulling="2025-12-01 08:37:53.291919506 +0000 UTC m=+1205.280977427" observedRunningTime="2025-12-01 08:37:53.93666524 +0000 UTC m=+1205.925723161" watchObservedRunningTime="2025-12-01 08:37:53.942764544 +0000 UTC m=+1205.931822465" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.382901 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.393717 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.396334 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.914593 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.952838 4744 generic.go:334] "Generic (PLEG): container finished" podID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerID="1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d" exitCode=0 Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.953399 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerDied","Data":"1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d"} Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.953581 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef069dd-c0d2-4235-b07c-e8ea6798d661","Type":"ContainerDied","Data":"43e4f5fd0801cc7d0ee852f6a879832f52d67ad5f932cf8343af7d003fbf9395"} Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.953613 4744 scope.go:117] "RemoveContainer" containerID="c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.953810 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.962625 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 08:37:55 crc kubenswrapper[4744]: I1201 08:37:55.993598 4744 scope.go:117] "RemoveContainer" containerID="2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.020123 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncdcc\" (UniqueName: \"kubernetes.io/projected/5ef069dd-c0d2-4235-b07c-e8ea6798d661-kube-api-access-ncdcc\") pod \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.020218 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-run-httpd\") pod \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.020304 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-log-httpd\") pod \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.020397 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-combined-ca-bundle\") pod \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.020445 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-config-data\") pod \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.020467 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-scripts\") pod \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.020532 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-sg-core-conf-yaml\") pod \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\" (UID: \"5ef069dd-c0d2-4235-b07c-e8ea6798d661\") " Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.021108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5ef069dd-c0d2-4235-b07c-e8ea6798d661" (UID: "5ef069dd-c0d2-4235-b07c-e8ea6798d661"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.021722 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.029774 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5ef069dd-c0d2-4235-b07c-e8ea6798d661" (UID: "5ef069dd-c0d2-4235-b07c-e8ea6798d661"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.029955 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-scripts" (OuterVolumeSpecName: "scripts") pod "5ef069dd-c0d2-4235-b07c-e8ea6798d661" (UID: "5ef069dd-c0d2-4235-b07c-e8ea6798d661"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.030228 4744 scope.go:117] "RemoveContainer" containerID="1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.038976 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef069dd-c0d2-4235-b07c-e8ea6798d661-kube-api-access-ncdcc" (OuterVolumeSpecName: "kube-api-access-ncdcc") pod "5ef069dd-c0d2-4235-b07c-e8ea6798d661" (UID: "5ef069dd-c0d2-4235-b07c-e8ea6798d661"). InnerVolumeSpecName "kube-api-access-ncdcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.057115 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5ef069dd-c0d2-4235-b07c-e8ea6798d661" (UID: "5ef069dd-c0d2-4235-b07c-e8ea6798d661"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.102965 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ef069dd-c0d2-4235-b07c-e8ea6798d661" (UID: "5ef069dd-c0d2-4235-b07c-e8ea6798d661"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.123783 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.124063 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.124074 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.124083 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncdcc\" (UniqueName: \"kubernetes.io/projected/5ef069dd-c0d2-4235-b07c-e8ea6798d661-kube-api-access-ncdcc\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.124111 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef069dd-c0d2-4235-b07c-e8ea6798d661-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.124912 4744 scope.go:117] "RemoveContainer" containerID="53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.127383 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-config-data" (OuterVolumeSpecName: "config-data") pod "5ef069dd-c0d2-4235-b07c-e8ea6798d661" (UID: "5ef069dd-c0d2-4235-b07c-e8ea6798d661"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.141948 4744 scope.go:117] "RemoveContainer" containerID="c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae" Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.142332 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae\": container with ID starting with c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae not found: ID does not exist" containerID="c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.142361 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae"} err="failed to get container status \"c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae\": rpc error: code = NotFound desc = could not find container \"c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae\": container with ID starting with c43b298106aa17cf7839cd85c7c58616d9829508125355aee63c7f6ea34042ae not found: ID does not exist" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.142383 4744 scope.go:117] "RemoveContainer" containerID="2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b" Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.142589 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b\": container with ID starting with 2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b not found: ID does not exist" containerID="2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.142673 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b"} err="failed to get container status \"2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b\": rpc error: code = NotFound desc = could not find container \"2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b\": container with ID starting with 2776efc49566ebdf09ad78f674a14c4f62d7d355da9275495e6be74ee5ddbc2b not found: ID does not exist" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.142741 4744 scope.go:117] "RemoveContainer" containerID="1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d" Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.143059 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d\": container with ID starting with 1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d not found: ID does not exist" containerID="1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.143108 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d"} err="failed to get container status \"1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d\": rpc error: code = NotFound desc = could not find container \"1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d\": container with ID starting with 1a8e631b9e2e254da20bfeaad123ff95a8990cb010b1a56d7cac6435b6630c7d not found: ID does not exist" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.143144 4744 scope.go:117] "RemoveContainer" containerID="53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0" Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.143438 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0\": container with ID starting with 53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0 not found: ID does not exist" containerID="53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.143527 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0"} err="failed to get container status \"53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0\": rpc error: code = NotFound desc = could not find container \"53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0\": container with ID starting with 53cea7d861da5a493c2608dc206b5c64b1ac3ca5dbd3dd6f36df29529ceef0f0 not found: ID does not exist" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.226051 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef069dd-c0d2-4235-b07c-e8ea6798d661-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.296046 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.297945 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.316643 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.317014 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="sg-core" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317029 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="sg-core" Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.317041 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="proxy-httpd" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317047 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="proxy-httpd" Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.317070 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-notification-agent" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317076 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-notification-agent" Dec 01 08:37:56 crc kubenswrapper[4744]: E1201 08:37:56.317086 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-central-agent" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317091 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-central-agent" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317265 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="proxy-httpd" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317285 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="sg-core" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317295 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-notification-agent" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.317305 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" containerName="ceilometer-central-agent" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.318827 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.321377 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.321635 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.321904 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.355804 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430511 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-run-httpd\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430675 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-scripts\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430714 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430748 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-log-httpd\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430795 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430892 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56ckl\" (UniqueName: \"kubernetes.io/projected/e63f6550-3248-4a38-9822-1bcb5ce8f720-kube-api-access-56ckl\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.430924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-config-data\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.552713 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.553856 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56ckl\" (UniqueName: \"kubernetes.io/projected/e63f6550-3248-4a38-9822-1bcb5ce8f720-kube-api-access-56ckl\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.553903 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-config-data\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.554078 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-run-httpd\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.554256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-scripts\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.554291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.554328 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-log-httpd\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.554365 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.554625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-run-httpd\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.554749 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-log-httpd\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.558290 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-scripts\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.560720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.560828 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.561101 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-config-data\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.561331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.570822 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56ckl\" (UniqueName: \"kubernetes.io/projected/e63f6550-3248-4a38-9822-1bcb5ce8f720-kube-api-access-56ckl\") pod \"ceilometer-0\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " pod="openstack/ceilometer-0" Dec 01 08:37:56 crc kubenswrapper[4744]: I1201 08:37:56.649148 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:37:57 crc kubenswrapper[4744]: I1201 08:37:57.099799 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:37:57 crc kubenswrapper[4744]: I1201 08:37:57.987603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerStarted","Data":"3430de011a0fd802d9f995388d59b3cfbeb960e7e1ebb2ecb4571fe450491626"} Dec 01 08:37:57 crc kubenswrapper[4744]: I1201 08:37:57.989303 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerStarted","Data":"0032f696f8dca935f179426c01c101772438d6db5b10861cb74a2affb13799c1"} Dec 01 08:37:58 crc kubenswrapper[4744]: I1201 08:37:58.303832 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef069dd-c0d2-4235-b07c-e8ea6798d661" path="/var/lib/kubelet/pods/5ef069dd-c0d2-4235-b07c-e8ea6798d661/volumes" Dec 01 08:37:58 crc kubenswrapper[4744]: I1201 08:37:58.926315 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:58 crc kubenswrapper[4744]: I1201 08:37:58.995709 4744 generic.go:334] "Generic (PLEG): container finished" podID="90e0478c-e863-4c2f-852c-673d676868c2" containerID="b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4" exitCode=137 Dec 01 08:37:58 crc kubenswrapper[4744]: I1201 08:37:58.995770 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"90e0478c-e863-4c2f-852c-673d676868c2","Type":"ContainerDied","Data":"b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4"} Dec 01 08:37:58 crc kubenswrapper[4744]: I1201 08:37:58.995800 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"90e0478c-e863-4c2f-852c-673d676868c2","Type":"ContainerDied","Data":"ccedb8b640c94510004f266c18a79c55269ca8f36af32aee9e6b6878c450b1fe"} Dec 01 08:37:58 crc kubenswrapper[4744]: I1201 08:37:58.995815 4744 scope.go:117] "RemoveContainer" containerID="b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4" Dec 01 08:37:58 crc kubenswrapper[4744]: I1201 08:37:58.995906 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.002452 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerStarted","Data":"bbcc6c36716297dccfa75594ab6ff83fcfda3a415f81e98acc1c8abaa09a28c8"} Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.022017 4744 scope.go:117] "RemoveContainer" containerID="b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4" Dec 01 08:37:59 crc kubenswrapper[4744]: E1201 08:37:59.022485 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4\": container with ID starting with b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4 not found: ID does not exist" containerID="b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.022519 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4"} err="failed to get container status \"b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4\": rpc error: code = NotFound desc = could not find container \"b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4\": container with ID starting with b615ab6109548833955842ec96c54580504cdec235831df334a8a07d891c74a4 not found: ID does not exist" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.110865 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-combined-ca-bundle\") pod \"90e0478c-e863-4c2f-852c-673d676868c2\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.110943 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twmg8\" (UniqueName: \"kubernetes.io/projected/90e0478c-e863-4c2f-852c-673d676868c2-kube-api-access-twmg8\") pod \"90e0478c-e863-4c2f-852c-673d676868c2\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.111012 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-config-data\") pod \"90e0478c-e863-4c2f-852c-673d676868c2\" (UID: \"90e0478c-e863-4c2f-852c-673d676868c2\") " Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.121699 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e0478c-e863-4c2f-852c-673d676868c2-kube-api-access-twmg8" (OuterVolumeSpecName: "kube-api-access-twmg8") pod "90e0478c-e863-4c2f-852c-673d676868c2" (UID: "90e0478c-e863-4c2f-852c-673d676868c2"). InnerVolumeSpecName "kube-api-access-twmg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.137276 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90e0478c-e863-4c2f-852c-673d676868c2" (UID: "90e0478c-e863-4c2f-852c-673d676868c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.139103 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-config-data" (OuterVolumeSpecName: "config-data") pod "90e0478c-e863-4c2f-852c-673d676868c2" (UID: "90e0478c-e863-4c2f-852c-673d676868c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.213263 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.213297 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twmg8\" (UniqueName: \"kubernetes.io/projected/90e0478c-e863-4c2f-852c-673d676868c2-kube-api-access-twmg8\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.213312 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e0478c-e863-4c2f-852c-673d676868c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.345322 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.367518 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.383478 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:59 crc kubenswrapper[4744]: E1201 08:37:59.384007 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e0478c-e863-4c2f-852c-673d676868c2" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.384030 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e0478c-e863-4c2f-852c-673d676868c2" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.384297 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e0478c-e863-4c2f-852c-673d676868c2" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.385300 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.388130 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.391163 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.391510 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.391702 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.517491 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.517573 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.517625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.517667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.517738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl42p\" (UniqueName: \"kubernetes.io/projected/c9f8476b-a1a9-42b5-ba40-95be5413d23b-kube-api-access-zl42p\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.619732 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.619824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.619899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.619949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.620045 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl42p\" (UniqueName: \"kubernetes.io/projected/c9f8476b-a1a9-42b5-ba40-95be5413d23b-kube-api-access-zl42p\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.624095 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.626590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.626626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.627571 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f8476b-a1a9-42b5-ba40-95be5413d23b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.638537 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl42p\" (UniqueName: \"kubernetes.io/projected/c9f8476b-a1a9-42b5-ba40-95be5413d23b-kube-api-access-zl42p\") pod \"nova-cell1-novncproxy-0\" (UID: \"c9f8476b-a1a9-42b5-ba40-95be5413d23b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:37:59 crc kubenswrapper[4744]: I1201 08:37:59.709915 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:38:00 crc kubenswrapper[4744]: I1201 08:38:00.012235 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerStarted","Data":"423dfabfabf975ad8141025cdeedcac055f7cdadb07ae18df32df9b11d73af2b"} Dec 01 08:38:00 crc kubenswrapper[4744]: I1201 08:38:00.174666 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 08:38:00 crc kubenswrapper[4744]: I1201 08:38:00.315456 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e0478c-e863-4c2f-852c-673d676868c2" path="/var/lib/kubelet/pods/90e0478c-e863-4c2f-852c-673d676868c2/volumes" Dec 01 08:38:01 crc kubenswrapper[4744]: I1201 08:38:01.028793 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c9f8476b-a1a9-42b5-ba40-95be5413d23b","Type":"ContainerStarted","Data":"42296f1a54121370a9957cd59073623f1d39a4026a2e8f7bf467481ab01263bf"} Dec 01 08:38:01 crc kubenswrapper[4744]: I1201 08:38:01.030126 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c9f8476b-a1a9-42b5-ba40-95be5413d23b","Type":"ContainerStarted","Data":"c069bee3b157a66060f974bfb2897aec66dee794bdee106944b21715aceaca3b"} Dec 01 08:38:01 crc kubenswrapper[4744]: I1201 08:38:01.051693 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.051668914 podStartE2EDuration="2.051668914s" podCreationTimestamp="2025-12-01 08:37:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:38:01.046331342 +0000 UTC m=+1213.035389273" watchObservedRunningTime="2025-12-01 08:38:01.051668914 +0000 UTC m=+1213.040726835" Dec 01 08:38:01 crc kubenswrapper[4744]: I1201 08:38:01.152549 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 08:38:01 crc kubenswrapper[4744]: I1201 08:38:01.153164 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 08:38:01 crc kubenswrapper[4744]: I1201 08:38:01.153659 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 08:38:01 crc kubenswrapper[4744]: I1201 08:38:01.156542 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.040115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerStarted","Data":"109cd186e5d781b460a5ff81eca298d0b97b87195d064b8c2c1469fba4b721c1"} Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.040285 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.044629 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.082368 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.394223696 podStartE2EDuration="6.08234545s" podCreationTimestamp="2025-12-01 08:37:56 +0000 UTC" firstStartedPulling="2025-12-01 08:37:57.113908079 +0000 UTC m=+1209.102966010" lastFinishedPulling="2025-12-01 08:38:00.802029843 +0000 UTC m=+1212.791087764" observedRunningTime="2025-12-01 08:38:02.068278889 +0000 UTC m=+1214.057336810" watchObservedRunningTime="2025-12-01 08:38:02.08234545 +0000 UTC m=+1214.071403371" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.236958 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-5q242"] Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.239253 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.255688 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-5q242"] Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.311682 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.371187 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.371330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.371387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.371433 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.371471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhtkx\" (UniqueName: \"kubernetes.io/projected/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-kube-api-access-jhtkx\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.371542 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-config\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.473105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-config\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.473324 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.473454 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.473485 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.473514 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.473545 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhtkx\" (UniqueName: \"kubernetes.io/projected/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-kube-api-access-jhtkx\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.474483 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.474499 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.478384 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.478449 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.479813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-config\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.492106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhtkx\" (UniqueName: \"kubernetes.io/projected/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-kube-api-access-jhtkx\") pod \"dnsmasq-dns-6d99f6bc7f-5q242\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:02 crc kubenswrapper[4744]: I1201 08:38:02.568559 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:03 crc kubenswrapper[4744]: I1201 08:38:03.048707 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:38:03 crc kubenswrapper[4744]: W1201 08:38:03.092330 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b4ff7ad_86db_42be_bb0f_9347c1c0494c.slice/crio-b1b0f3b0684c14178aacfb6c5ecbab017952ebf6b705bb6e382e4d97ec41266d WatchSource:0}: Error finding container b1b0f3b0684c14178aacfb6c5ecbab017952ebf6b705bb6e382e4d97ec41266d: Status 404 returned error can't find the container with id b1b0f3b0684c14178aacfb6c5ecbab017952ebf6b705bb6e382e4d97ec41266d Dec 01 08:38:03 crc kubenswrapper[4744]: I1201 08:38:03.098945 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-5q242"] Dec 01 08:38:04 crc kubenswrapper[4744]: I1201 08:38:04.068315 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerID="6f4ccadcd61c8ca2195b3c260389ade4801be0f481c699f658e28fe7d986cec2" exitCode=0 Dec 01 08:38:04 crc kubenswrapper[4744]: I1201 08:38:04.068459 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" event={"ID":"2b4ff7ad-86db-42be-bb0f-9347c1c0494c","Type":"ContainerDied","Data":"6f4ccadcd61c8ca2195b3c260389ade4801be0f481c699f658e28fe7d986cec2"} Dec 01 08:38:04 crc kubenswrapper[4744]: I1201 08:38:04.069846 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" event={"ID":"2b4ff7ad-86db-42be-bb0f-9347c1c0494c","Type":"ContainerStarted","Data":"b1b0f3b0684c14178aacfb6c5ecbab017952ebf6b705bb6e382e4d97ec41266d"} Dec 01 08:38:04 crc kubenswrapper[4744]: I1201 08:38:04.710683 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:38:04 crc kubenswrapper[4744]: I1201 08:38:04.748441 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.046934 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.082442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" event={"ID":"2b4ff7ad-86db-42be-bb0f-9347c1c0494c","Type":"ContainerStarted","Data":"e344853386235a5230a3c33ae9ce75dbe0ee43c71013f2662845ef0a9b12b195"} Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.082576 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-log" containerID="cri-o://b5629edf195a5d2c2e573937fb9b00785104f6e528f26e46fd3b40e0aed429a4" gracePeriod=30 Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.082656 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-api" containerID="cri-o://fc68715a13f1dc8c8706e40aa2f9d1f7c76b852168d83a09ac1025d5aed208a9" gracePeriod=30 Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.082853 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-central-agent" containerID="cri-o://3430de011a0fd802d9f995388d59b3cfbeb960e7e1ebb2ecb4571fe450491626" gracePeriod=30 Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.082901 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="proxy-httpd" containerID="cri-o://109cd186e5d781b460a5ff81eca298d0b97b87195d064b8c2c1469fba4b721c1" gracePeriod=30 Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.082949 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="sg-core" containerID="cri-o://423dfabfabf975ad8141025cdeedcac055f7cdadb07ae18df32df9b11d73af2b" gracePeriod=30 Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.082945 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-notification-agent" containerID="cri-o://bbcc6c36716297dccfa75594ab6ff83fcfda3a415f81e98acc1c8abaa09a28c8" gracePeriod=30 Dec 01 08:38:05 crc kubenswrapper[4744]: I1201 08:38:05.116940 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" podStartSLOduration=3.116916817 podStartE2EDuration="3.116916817s" podCreationTimestamp="2025-12-01 08:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:38:05.11070483 +0000 UTC m=+1217.099762751" watchObservedRunningTime="2025-12-01 08:38:05.116916817 +0000 UTC m=+1217.105974738" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.092265 4744 generic.go:334] "Generic (PLEG): container finished" podID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerID="b5629edf195a5d2c2e573937fb9b00785104f6e528f26e46fd3b40e0aed429a4" exitCode=143 Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.092314 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d74fdf4b-20a8-4df0-96c0-ff5e06339169","Type":"ContainerDied","Data":"b5629edf195a5d2c2e573937fb9b00785104f6e528f26e46fd3b40e0aed429a4"} Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094687 4744 generic.go:334] "Generic (PLEG): container finished" podID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerID="109cd186e5d781b460a5ff81eca298d0b97b87195d064b8c2c1469fba4b721c1" exitCode=0 Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094714 4744 generic.go:334] "Generic (PLEG): container finished" podID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerID="423dfabfabf975ad8141025cdeedcac055f7cdadb07ae18df32df9b11d73af2b" exitCode=2 Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094724 4744 generic.go:334] "Generic (PLEG): container finished" podID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerID="bbcc6c36716297dccfa75594ab6ff83fcfda3a415f81e98acc1c8abaa09a28c8" exitCode=0 Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094733 4744 generic.go:334] "Generic (PLEG): container finished" podID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerID="3430de011a0fd802d9f995388d59b3cfbeb960e7e1ebb2ecb4571fe450491626" exitCode=0 Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094761 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerDied","Data":"109cd186e5d781b460a5ff81eca298d0b97b87195d064b8c2c1469fba4b721c1"} Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094795 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerDied","Data":"423dfabfabf975ad8141025cdeedcac055f7cdadb07ae18df32df9b11d73af2b"} Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerDied","Data":"bbcc6c36716297dccfa75594ab6ff83fcfda3a415f81e98acc1c8abaa09a28c8"} Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.094819 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerDied","Data":"3430de011a0fd802d9f995388d59b3cfbeb960e7e1ebb2ecb4571fe450491626"} Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.095033 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.667082 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-sg-core-conf-yaml\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-ceilometer-tls-certs\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763226 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56ckl\" (UniqueName: \"kubernetes.io/projected/e63f6550-3248-4a38-9822-1bcb5ce8f720-kube-api-access-56ckl\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-config-data\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763324 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-combined-ca-bundle\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763372 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-scripts\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763496 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-log-httpd\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.763546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-run-httpd\") pod \"e63f6550-3248-4a38-9822-1bcb5ce8f720\" (UID: \"e63f6550-3248-4a38-9822-1bcb5ce8f720\") " Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.764199 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.764458 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.769726 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-scripts" (OuterVolumeSpecName: "scripts") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.770146 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e63f6550-3248-4a38-9822-1bcb5ce8f720-kube-api-access-56ckl" (OuterVolumeSpecName: "kube-api-access-56ckl") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "kube-api-access-56ckl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.803700 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.831375 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.865981 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.866042 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.866059 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56ckl\" (UniqueName: \"kubernetes.io/projected/e63f6550-3248-4a38-9822-1bcb5ce8f720-kube-api-access-56ckl\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.866076 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.866089 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.866100 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e63f6550-3248-4a38-9822-1bcb5ce8f720-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.882020 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.886174 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-config-data" (OuterVolumeSpecName: "config-data") pod "e63f6550-3248-4a38-9822-1bcb5ce8f720" (UID: "e63f6550-3248-4a38-9822-1bcb5ce8f720"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.967848 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:06 crc kubenswrapper[4744]: I1201 08:38:06.967883 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e63f6550-3248-4a38-9822-1bcb5ce8f720-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.106294 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.106296 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e63f6550-3248-4a38-9822-1bcb5ce8f720","Type":"ContainerDied","Data":"0032f696f8dca935f179426c01c101772438d6db5b10861cb74a2affb13799c1"} Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.106368 4744 scope.go:117] "RemoveContainer" containerID="109cd186e5d781b460a5ff81eca298d0b97b87195d064b8c2c1469fba4b721c1" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.141269 4744 scope.go:117] "RemoveContainer" containerID="423dfabfabf975ad8141025cdeedcac055f7cdadb07ae18df32df9b11d73af2b" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.163906 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.184037 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.185197 4744 scope.go:117] "RemoveContainer" containerID="bbcc6c36716297dccfa75594ab6ff83fcfda3a415f81e98acc1c8abaa09a28c8" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.210955 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:38:07 crc kubenswrapper[4744]: E1201 08:38:07.211443 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="proxy-httpd" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.211468 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="proxy-httpd" Dec 01 08:38:07 crc kubenswrapper[4744]: E1201 08:38:07.211502 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-central-agent" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.211511 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-central-agent" Dec 01 08:38:07 crc kubenswrapper[4744]: E1201 08:38:07.211534 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="sg-core" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.211542 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="sg-core" Dec 01 08:38:07 crc kubenswrapper[4744]: E1201 08:38:07.211569 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-notification-agent" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.211578 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-notification-agent" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.212143 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="sg-core" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.212171 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="proxy-httpd" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.212189 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-notification-agent" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.212207 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" containerName="ceilometer-central-agent" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.220136 4744 scope.go:117] "RemoveContainer" containerID="3430de011a0fd802d9f995388d59b3cfbeb960e7e1ebb2ecb4571fe450491626" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.229611 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.233217 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.233241 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.233657 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.240907 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.377999 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.378068 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.378143 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-run-httpd\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.378167 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-scripts\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.378312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-log-httpd\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.378356 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.378458 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-config-data\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.378523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rklpx\" (UniqueName: \"kubernetes.io/projected/340586bd-ec9d-4a58-a46d-1d043557b994-kube-api-access-rklpx\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.480386 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.480479 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.480588 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-run-httpd\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.480625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-scripts\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.480685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-log-httpd\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.481136 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.481163 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-run-httpd\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.481245 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-config-data\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.481303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rklpx\" (UniqueName: \"kubernetes.io/projected/340586bd-ec9d-4a58-a46d-1d043557b994-kube-api-access-rklpx\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.481305 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-log-httpd\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.484390 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.484711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-scripts\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.485078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.485589 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-config-data\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.485813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.498217 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rklpx\" (UniqueName: \"kubernetes.io/projected/340586bd-ec9d-4a58-a46d-1d043557b994-kube-api-access-rklpx\") pod \"ceilometer-0\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " pod="openstack/ceilometer-0" Dec 01 08:38:07 crc kubenswrapper[4744]: I1201 08:38:07.559790 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 08:38:08 crc kubenswrapper[4744]: W1201 08:38:08.039910 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod340586bd_ec9d_4a58_a46d_1d043557b994.slice/crio-75ac7023db2237a9400f179e7bf2ae916b06dcca79cba17a9a4fcd2ca5fd3bc0 WatchSource:0}: Error finding container 75ac7023db2237a9400f179e7bf2ae916b06dcca79cba17a9a4fcd2ca5fd3bc0: Status 404 returned error can't find the container with id 75ac7023db2237a9400f179e7bf2ae916b06dcca79cba17a9a4fcd2ca5fd3bc0 Dec 01 08:38:08 crc kubenswrapper[4744]: I1201 08:38:08.040391 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 08:38:08 crc kubenswrapper[4744]: I1201 08:38:08.116664 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerStarted","Data":"75ac7023db2237a9400f179e7bf2ae916b06dcca79cba17a9a4fcd2ca5fd3bc0"} Dec 01 08:38:08 crc kubenswrapper[4744]: I1201 08:38:08.298138 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e63f6550-3248-4a38-9822-1bcb5ce8f720" path="/var/lib/kubelet/pods/e63f6550-3248-4a38-9822-1bcb5ce8f720/volumes" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.128032 4744 generic.go:334] "Generic (PLEG): container finished" podID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerID="fc68715a13f1dc8c8706e40aa2f9d1f7c76b852168d83a09ac1025d5aed208a9" exitCode=0 Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.128108 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d74fdf4b-20a8-4df0-96c0-ff5e06339169","Type":"ContainerDied","Data":"fc68715a13f1dc8c8706e40aa2f9d1f7c76b852168d83a09ac1025d5aed208a9"} Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.130118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerStarted","Data":"eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed"} Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.299838 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.422966 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-combined-ca-bundle\") pod \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.423024 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74fdf4b-20a8-4df0-96c0-ff5e06339169-logs\") pod \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.423233 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfh24\" (UniqueName: \"kubernetes.io/projected/d74fdf4b-20a8-4df0-96c0-ff5e06339169-kube-api-access-tfh24\") pod \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.423278 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-config-data\") pod \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\" (UID: \"d74fdf4b-20a8-4df0-96c0-ff5e06339169\") " Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.423668 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d74fdf4b-20a8-4df0-96c0-ff5e06339169-logs" (OuterVolumeSpecName: "logs") pod "d74fdf4b-20a8-4df0-96c0-ff5e06339169" (UID: "d74fdf4b-20a8-4df0-96c0-ff5e06339169"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.423993 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74fdf4b-20a8-4df0-96c0-ff5e06339169-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.439732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d74fdf4b-20a8-4df0-96c0-ff5e06339169-kube-api-access-tfh24" (OuterVolumeSpecName: "kube-api-access-tfh24") pod "d74fdf4b-20a8-4df0-96c0-ff5e06339169" (UID: "d74fdf4b-20a8-4df0-96c0-ff5e06339169"). InnerVolumeSpecName "kube-api-access-tfh24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.453611 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d74fdf4b-20a8-4df0-96c0-ff5e06339169" (UID: "d74fdf4b-20a8-4df0-96c0-ff5e06339169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.480772 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-config-data" (OuterVolumeSpecName: "config-data") pod "d74fdf4b-20a8-4df0-96c0-ff5e06339169" (UID: "d74fdf4b-20a8-4df0-96c0-ff5e06339169"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.528056 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.528103 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfh24\" (UniqueName: \"kubernetes.io/projected/d74fdf4b-20a8-4df0-96c0-ff5e06339169-kube-api-access-tfh24\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.528122 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74fdf4b-20a8-4df0-96c0-ff5e06339169-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.710756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:38:09 crc kubenswrapper[4744]: I1201 08:38:09.739689 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.141622 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d74fdf4b-20a8-4df0-96c0-ff5e06339169","Type":"ContainerDied","Data":"997577237a5d66040871dfb92561ab5651a401e12cb7ace550d091f55bbfeee8"} Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.141880 4744 scope.go:117] "RemoveContainer" containerID="fc68715a13f1dc8c8706e40aa2f9d1f7c76b852168d83a09ac1025d5aed208a9" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.141957 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.144631 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerStarted","Data":"2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d"} Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.164711 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.166101 4744 scope.go:117] "RemoveContainer" containerID="b5629edf195a5d2c2e573937fb9b00785104f6e528f26e46fd3b40e0aed429a4" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.194539 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.204501 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.222653 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:10 crc kubenswrapper[4744]: E1201 08:38:10.223831 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-log" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.223867 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-log" Dec 01 08:38:10 crc kubenswrapper[4744]: E1201 08:38:10.223894 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-api" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.223902 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-api" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.224098 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-api" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.224128 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" containerName="nova-api-log" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.225249 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.235306 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.236009 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.236256 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.236559 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.244711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-internal-tls-certs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.244791 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-config-data\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.244902 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.244927 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70c618c0-c876-450e-98b1-e4e562851407-logs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.244946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-public-tls-certs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.245038 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9g47\" (UniqueName: \"kubernetes.io/projected/70c618c0-c876-450e-98b1-e4e562851407-kube-api-access-c9g47\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.296288 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d74fdf4b-20a8-4df0-96c0-ff5e06339169" path="/var/lib/kubelet/pods/d74fdf4b-20a8-4df0-96c0-ff5e06339169/volumes" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.346983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.347027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70c618c0-c876-450e-98b1-e4e562851407-logs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.347046 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-public-tls-certs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.347126 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9g47\" (UniqueName: \"kubernetes.io/projected/70c618c0-c876-450e-98b1-e4e562851407-kube-api-access-c9g47\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.347206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-internal-tls-certs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.347237 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-config-data\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.348096 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70c618c0-c876-450e-98b1-e4e562851407-logs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.351157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-config-data\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.353368 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-internal-tls-certs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.355785 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-public-tls-certs\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.360805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.370808 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9g47\" (UniqueName: \"kubernetes.io/projected/70c618c0-c876-450e-98b1-e4e562851407-kube-api-access-c9g47\") pod \"nova-api-0\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.407358 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-klrlq"] Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.409127 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.411956 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.412160 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.418576 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-klrlq"] Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.448459 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.448518 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c897l\" (UniqueName: \"kubernetes.io/projected/a49df21a-3b70-4797-8501-516a2f39e4f2-kube-api-access-c897l\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.448590 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-scripts\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.448650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-config-data\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.548375 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.549498 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c897l\" (UniqueName: \"kubernetes.io/projected/a49df21a-3b70-4797-8501-516a2f39e4f2-kube-api-access-c897l\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.549562 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-scripts\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.549611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-config-data\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.549689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.554641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-scripts\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.557465 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-config-data\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.557856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.569832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c897l\" (UniqueName: \"kubernetes.io/projected/a49df21a-3b70-4797-8501-516a2f39e4f2-kube-api-access-c897l\") pod \"nova-cell1-cell-mapping-klrlq\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:10 crc kubenswrapper[4744]: I1201 08:38:10.777220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:11 crc kubenswrapper[4744]: I1201 08:38:11.056480 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:11 crc kubenswrapper[4744]: I1201 08:38:11.155124 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerStarted","Data":"91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd"} Dec 01 08:38:11 crc kubenswrapper[4744]: I1201 08:38:11.156852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70c618c0-c876-450e-98b1-e4e562851407","Type":"ContainerStarted","Data":"9a4bff47642322e02b77d57b5fc75b46d5aa0cf8adb5e2a5351d0533032d4a9a"} Dec 01 08:38:11 crc kubenswrapper[4744]: I1201 08:38:11.250330 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-klrlq"] Dec 01 08:38:11 crc kubenswrapper[4744]: W1201 08:38:11.257196 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda49df21a_3b70_4797_8501_516a2f39e4f2.slice/crio-f140cbd80e661b961e44cae1140537d6746be511f31d3bfeca4fb5018f996dbe WatchSource:0}: Error finding container f140cbd80e661b961e44cae1140537d6746be511f31d3bfeca4fb5018f996dbe: Status 404 returned error can't find the container with id f140cbd80e661b961e44cae1140537d6746be511f31d3bfeca4fb5018f996dbe Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.169261 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-klrlq" event={"ID":"a49df21a-3b70-4797-8501-516a2f39e4f2","Type":"ContainerStarted","Data":"ec9a69156d333c2d20091a2617c10676d492dbcd7b26f3315c419be270edbd4c"} Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.170128 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-klrlq" event={"ID":"a49df21a-3b70-4797-8501-516a2f39e4f2","Type":"ContainerStarted","Data":"f140cbd80e661b961e44cae1140537d6746be511f31d3bfeca4fb5018f996dbe"} Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.171918 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70c618c0-c876-450e-98b1-e4e562851407","Type":"ContainerStarted","Data":"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af"} Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.171971 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70c618c0-c876-450e-98b1-e4e562851407","Type":"ContainerStarted","Data":"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c"} Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.193209 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-klrlq" podStartSLOduration=2.1931909960000002 podStartE2EDuration="2.193190996s" podCreationTimestamp="2025-12-01 08:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:38:12.188125052 +0000 UTC m=+1224.177182973" watchObservedRunningTime="2025-12-01 08:38:12.193190996 +0000 UTC m=+1224.182248917" Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.569695 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.587834 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.587815515 podStartE2EDuration="2.587815515s" podCreationTimestamp="2025-12-01 08:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:38:12.205142607 +0000 UTC m=+1224.194200528" watchObservedRunningTime="2025-12-01 08:38:12.587815515 +0000 UTC m=+1224.576873436" Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.640668 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-cm67w"] Dec 01 08:38:12 crc kubenswrapper[4744]: I1201 08:38:12.640969 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7877d89589-cm67w" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerName="dnsmasq-dns" containerID="cri-o://f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2" gracePeriod=10 Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.175868 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.184203 4744 generic.go:334] "Generic (PLEG): container finished" podID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerID="f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2" exitCode=0 Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.184264 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-cm67w" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.184302 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-cm67w" event={"ID":"5129f5b1-84f9-4c6c-95ec-c0a8396ac082","Type":"ContainerDied","Data":"f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2"} Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.184356 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-cm67w" event={"ID":"5129f5b1-84f9-4c6c-95ec-c0a8396ac082","Type":"ContainerDied","Data":"8a27b689c67fe164c070c4a7a662b88d83b3172a829112bc783505ab33d7cf33"} Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.184379 4744 scope.go:117] "RemoveContainer" containerID="f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.188258 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerStarted","Data":"a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06"} Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.201905 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-sb\") pod \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.201975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-nb\") pod \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.202045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-config\") pod \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.202074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-swift-storage-0\") pod \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.202116 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsjdd\" (UniqueName: \"kubernetes.io/projected/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-kube-api-access-dsjdd\") pod \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.202139 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-svc\") pod \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\" (UID: \"5129f5b1-84f9-4c6c-95ec-c0a8396ac082\") " Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.210532 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-kube-api-access-dsjdd" (OuterVolumeSpecName: "kube-api-access-dsjdd") pod "5129f5b1-84f9-4c6c-95ec-c0a8396ac082" (UID: "5129f5b1-84f9-4c6c-95ec-c0a8396ac082"). InnerVolumeSpecName "kube-api-access-dsjdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.228002 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.182813199 podStartE2EDuration="6.22798218s" podCreationTimestamp="2025-12-01 08:38:07 +0000 UTC" firstStartedPulling="2025-12-01 08:38:08.042600138 +0000 UTC m=+1220.031658069" lastFinishedPulling="2025-12-01 08:38:12.087769129 +0000 UTC m=+1224.076827050" observedRunningTime="2025-12-01 08:38:13.213776524 +0000 UTC m=+1225.202834445" watchObservedRunningTime="2025-12-01 08:38:13.22798218 +0000 UTC m=+1225.217040101" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.233162 4744 scope.go:117] "RemoveContainer" containerID="247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.285506 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5129f5b1-84f9-4c6c-95ec-c0a8396ac082" (UID: "5129f5b1-84f9-4c6c-95ec-c0a8396ac082"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.287589 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5129f5b1-84f9-4c6c-95ec-c0a8396ac082" (UID: "5129f5b1-84f9-4c6c-95ec-c0a8396ac082"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.287719 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5129f5b1-84f9-4c6c-95ec-c0a8396ac082" (UID: "5129f5b1-84f9-4c6c-95ec-c0a8396ac082"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.304603 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.304635 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.304645 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsjdd\" (UniqueName: \"kubernetes.io/projected/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-kube-api-access-dsjdd\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.304654 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.306456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-config" (OuterVolumeSpecName: "config") pod "5129f5b1-84f9-4c6c-95ec-c0a8396ac082" (UID: "5129f5b1-84f9-4c6c-95ec-c0a8396ac082"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.335124 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5129f5b1-84f9-4c6c-95ec-c0a8396ac082" (UID: "5129f5b1-84f9-4c6c-95ec-c0a8396ac082"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.404975 4744 scope.go:117] "RemoveContainer" containerID="f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2" Dec 01 08:38:13 crc kubenswrapper[4744]: E1201 08:38:13.405514 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2\": container with ID starting with f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2 not found: ID does not exist" containerID="f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.405551 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2"} err="failed to get container status \"f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2\": rpc error: code = NotFound desc = could not find container \"f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2\": container with ID starting with f055c53187a72e4d56c6dddb309c13e4b9496b7bcedab66b445be21e1bfd3ef2 not found: ID does not exist" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.405577 4744 scope.go:117] "RemoveContainer" containerID="247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0" Dec 01 08:38:13 crc kubenswrapper[4744]: E1201 08:38:13.405866 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0\": container with ID starting with 247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0 not found: ID does not exist" containerID="247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.405893 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0"} err="failed to get container status \"247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0\": rpc error: code = NotFound desc = could not find container \"247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0\": container with ID starting with 247e14d9a1a1520c8ea3ad18a5404d676023c0c9c3632eb821ee26252bcf66c0 not found: ID does not exist" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.406324 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.406345 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5129f5b1-84f9-4c6c-95ec-c0a8396ac082-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.538685 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-cm67w"] Dec 01 08:38:13 crc kubenswrapper[4744]: I1201 08:38:13.550047 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-cm67w"] Dec 01 08:38:14 crc kubenswrapper[4744]: I1201 08:38:14.201292 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 08:38:14 crc kubenswrapper[4744]: I1201 08:38:14.296239 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" path="/var/lib/kubelet/pods/5129f5b1-84f9-4c6c-95ec-c0a8396ac082/volumes" Dec 01 08:38:16 crc kubenswrapper[4744]: I1201 08:38:16.229115 4744 generic.go:334] "Generic (PLEG): container finished" podID="a49df21a-3b70-4797-8501-516a2f39e4f2" containerID="ec9a69156d333c2d20091a2617c10676d492dbcd7b26f3315c419be270edbd4c" exitCode=0 Dec 01 08:38:16 crc kubenswrapper[4744]: I1201 08:38:16.229488 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-klrlq" event={"ID":"a49df21a-3b70-4797-8501-516a2f39e4f2","Type":"ContainerDied","Data":"ec9a69156d333c2d20091a2617c10676d492dbcd7b26f3315c419be270edbd4c"} Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.715578 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.894827 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c897l\" (UniqueName: \"kubernetes.io/projected/a49df21a-3b70-4797-8501-516a2f39e4f2-kube-api-access-c897l\") pod \"a49df21a-3b70-4797-8501-516a2f39e4f2\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.894905 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-scripts\") pod \"a49df21a-3b70-4797-8501-516a2f39e4f2\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.895018 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-config-data\") pod \"a49df21a-3b70-4797-8501-516a2f39e4f2\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.895062 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-combined-ca-bundle\") pod \"a49df21a-3b70-4797-8501-516a2f39e4f2\" (UID: \"a49df21a-3b70-4797-8501-516a2f39e4f2\") " Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.901115 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a49df21a-3b70-4797-8501-516a2f39e4f2-kube-api-access-c897l" (OuterVolumeSpecName: "kube-api-access-c897l") pod "a49df21a-3b70-4797-8501-516a2f39e4f2" (UID: "a49df21a-3b70-4797-8501-516a2f39e4f2"). InnerVolumeSpecName "kube-api-access-c897l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.903728 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-scripts" (OuterVolumeSpecName: "scripts") pod "a49df21a-3b70-4797-8501-516a2f39e4f2" (UID: "a49df21a-3b70-4797-8501-516a2f39e4f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.930468 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-config-data" (OuterVolumeSpecName: "config-data") pod "a49df21a-3b70-4797-8501-516a2f39e4f2" (UID: "a49df21a-3b70-4797-8501-516a2f39e4f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.946171 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a49df21a-3b70-4797-8501-516a2f39e4f2" (UID: "a49df21a-3b70-4797-8501-516a2f39e4f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.997873 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c897l\" (UniqueName: \"kubernetes.io/projected/a49df21a-3b70-4797-8501-516a2f39e4f2-kube-api-access-c897l\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.997939 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.997950 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:17 crc kubenswrapper[4744]: I1201 08:38:17.997962 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49df21a-3b70-4797-8501-516a2f39e4f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.071310 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7877d89589-cm67w" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.197:5353: i/o timeout" Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.258005 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-klrlq" event={"ID":"a49df21a-3b70-4797-8501-516a2f39e4f2","Type":"ContainerDied","Data":"f140cbd80e661b961e44cae1140537d6746be511f31d3bfeca4fb5018f996dbe"} Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.258060 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-klrlq" Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.258065 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f140cbd80e661b961e44cae1140537d6746be511f31d3bfeca4fb5018f996dbe" Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.460470 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.460792 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-log" containerID="cri-o://b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c" gracePeriod=30 Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.460825 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-api" containerID="cri-o://3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af" gracePeriod=30 Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.473685 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.474048 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8bb988e1-ef1a-45b8-b3fc-cf6025835a89" containerName="nova-scheduler-scheduler" containerID="cri-o://f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb" gracePeriod=30 Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.515724 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.515961 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-log" containerID="cri-o://b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059" gracePeriod=30 Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.516049 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-metadata" containerID="cri-o://9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd" gracePeriod=30 Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.633919 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.633975 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.634016 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.634619 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e4f0e7898cde7a29868d7677f91a6fae395c27e58d4476c4df649466e26b91b"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:38:18 crc kubenswrapper[4744]: I1201 08:38:18.634683 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://1e4f0e7898cde7a29868d7677f91a6fae395c27e58d4476c4df649466e26b91b" gracePeriod=600 Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.057470 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.223561 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-internal-tls-certs\") pod \"70c618c0-c876-450e-98b1-e4e562851407\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.224075 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-combined-ca-bundle\") pod \"70c618c0-c876-450e-98b1-e4e562851407\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.224126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70c618c0-c876-450e-98b1-e4e562851407-logs\") pod \"70c618c0-c876-450e-98b1-e4e562851407\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.224171 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-config-data\") pod \"70c618c0-c876-450e-98b1-e4e562851407\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.224205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9g47\" (UniqueName: \"kubernetes.io/projected/70c618c0-c876-450e-98b1-e4e562851407-kube-api-access-c9g47\") pod \"70c618c0-c876-450e-98b1-e4e562851407\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.224271 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-public-tls-certs\") pod \"70c618c0-c876-450e-98b1-e4e562851407\" (UID: \"70c618c0-c876-450e-98b1-e4e562851407\") " Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.224938 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70c618c0-c876-450e-98b1-e4e562851407-logs" (OuterVolumeSpecName: "logs") pod "70c618c0-c876-450e-98b1-e4e562851407" (UID: "70c618c0-c876-450e-98b1-e4e562851407"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.230842 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c618c0-c876-450e-98b1-e4e562851407-kube-api-access-c9g47" (OuterVolumeSpecName: "kube-api-access-c9g47") pod "70c618c0-c876-450e-98b1-e4e562851407" (UID: "70c618c0-c876-450e-98b1-e4e562851407"). InnerVolumeSpecName "kube-api-access-c9g47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.256002 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70c618c0-c876-450e-98b1-e4e562851407" (UID: "70c618c0-c876-450e-98b1-e4e562851407"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.257951 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-config-data" (OuterVolumeSpecName: "config-data") pod "70c618c0-c876-450e-98b1-e4e562851407" (UID: "70c618c0-c876-450e-98b1-e4e562851407"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.268899 4744 generic.go:334] "Generic (PLEG): container finished" podID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerID="b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059" exitCode=143 Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.268970 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19dbb279-f666-4a27-a699-88ccd0b5cf2b","Type":"ContainerDied","Data":"b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059"} Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.271675 4744 generic.go:334] "Generic (PLEG): container finished" podID="70c618c0-c876-450e-98b1-e4e562851407" containerID="3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af" exitCode=0 Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.271736 4744 generic.go:334] "Generic (PLEG): container finished" podID="70c618c0-c876-450e-98b1-e4e562851407" containerID="b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c" exitCode=143 Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.271777 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.271757 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70c618c0-c876-450e-98b1-e4e562851407","Type":"ContainerDied","Data":"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af"} Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.271906 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70c618c0-c876-450e-98b1-e4e562851407","Type":"ContainerDied","Data":"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c"} Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.271921 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70c618c0-c876-450e-98b1-e4e562851407","Type":"ContainerDied","Data":"9a4bff47642322e02b77d57b5fc75b46d5aa0cf8adb5e2a5351d0533032d4a9a"} Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.271940 4744 scope.go:117] "RemoveContainer" containerID="3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.275777 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="1e4f0e7898cde7a29868d7677f91a6fae395c27e58d4476c4df649466e26b91b" exitCode=0 Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.275833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"1e4f0e7898cde7a29868d7677f91a6fae395c27e58d4476c4df649466e26b91b"} Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.275880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"610c82f121426bae487451dc7fd1be93ad3b6ec0bd33a7499656278834e982ab"} Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.290062 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "70c618c0-c876-450e-98b1-e4e562851407" (UID: "70c618c0-c876-450e-98b1-e4e562851407"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.299232 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "70c618c0-c876-450e-98b1-e4e562851407" (UID: "70c618c0-c876-450e-98b1-e4e562851407"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.312582 4744 scope.go:117] "RemoveContainer" containerID="b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.327195 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.327233 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.327242 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.327250 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70c618c0-c876-450e-98b1-e4e562851407-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.327259 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c618c0-c876-450e-98b1-e4e562851407-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.327267 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9g47\" (UniqueName: \"kubernetes.io/projected/70c618c0-c876-450e-98b1-e4e562851407-kube-api-access-c9g47\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.333697 4744 scope.go:117] "RemoveContainer" containerID="3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af" Dec 01 08:38:19 crc kubenswrapper[4744]: E1201 08:38:19.334187 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af\": container with ID starting with 3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af not found: ID does not exist" containerID="3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.334234 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af"} err="failed to get container status \"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af\": rpc error: code = NotFound desc = could not find container \"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af\": container with ID starting with 3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af not found: ID does not exist" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.334267 4744 scope.go:117] "RemoveContainer" containerID="b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c" Dec 01 08:38:19 crc kubenswrapper[4744]: E1201 08:38:19.334620 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c\": container with ID starting with b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c not found: ID does not exist" containerID="b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.334658 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c"} err="failed to get container status \"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c\": rpc error: code = NotFound desc = could not find container \"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c\": container with ID starting with b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c not found: ID does not exist" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.334683 4744 scope.go:117] "RemoveContainer" containerID="3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.345811 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af"} err="failed to get container status \"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af\": rpc error: code = NotFound desc = could not find container \"3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af\": container with ID starting with 3bf05641a66e5f6e6ab7e76c5fa8941db73ae255c9c454925bb52912d4e358af not found: ID does not exist" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.345858 4744 scope.go:117] "RemoveContainer" containerID="b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.346140 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c"} err="failed to get container status \"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c\": rpc error: code = NotFound desc = could not find container \"b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c\": container with ID starting with b097a54b7d7ffb16193941128ef854698055d8f600d449ab7b67061c45bf497c not found: ID does not exist" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.346161 4744 scope.go:117] "RemoveContainer" containerID="64ec6b8815aa74952cf65128020a2c3112aeb4c1255a0d0fff478a4f4a871f94" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.616780 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.649091 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.667239 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:19 crc kubenswrapper[4744]: E1201 08:38:19.667688 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-api" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.667728 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-api" Dec 01 08:38:19 crc kubenswrapper[4744]: E1201 08:38:19.667742 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerName="init" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.667749 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerName="init" Dec 01 08:38:19 crc kubenswrapper[4744]: E1201 08:38:19.667766 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a49df21a-3b70-4797-8501-516a2f39e4f2" containerName="nova-manage" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.667772 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a49df21a-3b70-4797-8501-516a2f39e4f2" containerName="nova-manage" Dec 01 08:38:19 crc kubenswrapper[4744]: E1201 08:38:19.667788 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerName="dnsmasq-dns" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.667794 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerName="dnsmasq-dns" Dec 01 08:38:19 crc kubenswrapper[4744]: E1201 08:38:19.667811 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-log" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.667817 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-log" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.667989 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5129f5b1-84f9-4c6c-95ec-c0a8396ac082" containerName="dnsmasq-dns" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.668003 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-api" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.668019 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c618c0-c876-450e-98b1-e4e562851407" containerName="nova-api-log" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.668029 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a49df21a-3b70-4797-8501-516a2f39e4f2" containerName="nova-manage" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.669042 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.671574 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.671943 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.674350 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.684761 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.836693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec302fdc-d505-4f9c-b403-14643c501988-logs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.836741 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-public-tls-certs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.837169 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.837293 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-config-data\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.837401 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvkp4\" (UniqueName: \"kubernetes.io/projected/ec302fdc-d505-4f9c-b403-14643c501988-kube-api-access-bvkp4\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.837564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.938861 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.939241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-config-data\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.939277 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvkp4\" (UniqueName: \"kubernetes.io/projected/ec302fdc-d505-4f9c-b403-14643c501988-kube-api-access-bvkp4\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.939313 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.939370 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec302fdc-d505-4f9c-b403-14643c501988-logs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.939389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-public-tls-certs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.940310 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec302fdc-d505-4f9c-b403-14643c501988-logs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.944333 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-config-data\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.944357 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.946177 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.946745 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec302fdc-d505-4f9c-b403-14643c501988-public-tls-certs\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:19 crc kubenswrapper[4744]: I1201 08:38:19.959361 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvkp4\" (UniqueName: \"kubernetes.io/projected/ec302fdc-d505-4f9c-b403-14643c501988-kube-api-access-bvkp4\") pod \"nova-api-0\" (UID: \"ec302fdc-d505-4f9c-b403-14643c501988\") " pod="openstack/nova-api-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.060421 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.199144 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.288095 4744 generic.go:334] "Generic (PLEG): container finished" podID="8bb988e1-ef1a-45b8-b3fc-cf6025835a89" containerID="f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb" exitCode=0 Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.288170 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.297646 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c618c0-c876-450e-98b1-e4e562851407" path="/var/lib/kubelet/pods/70c618c0-c876-450e-98b1-e4e562851407/volumes" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.301327 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bb988e1-ef1a-45b8-b3fc-cf6025835a89","Type":"ContainerDied","Data":"f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb"} Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.301373 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bb988e1-ef1a-45b8-b3fc-cf6025835a89","Type":"ContainerDied","Data":"94c07d76039e5dd41953446ecbeceb3f7a3f6d3b5fb95da5d1ce0e80d166f766"} Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.301391 4744 scope.go:117] "RemoveContainer" containerID="f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.327476 4744 scope.go:117] "RemoveContainer" containerID="f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb" Dec 01 08:38:20 crc kubenswrapper[4744]: E1201 08:38:20.327992 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb\": container with ID starting with f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb not found: ID does not exist" containerID="f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.328032 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb"} err="failed to get container status \"f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb\": rpc error: code = NotFound desc = could not find container \"f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb\": container with ID starting with f0fc2f5400e449205f1c4ce8d77e8991f4480022e4d7e9e151eea27b71e606fb not found: ID does not exist" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.347777 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-config-data\") pod \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.347829 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-combined-ca-bundle\") pod \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.347937 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmbq9\" (UniqueName: \"kubernetes.io/projected/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-kube-api-access-pmbq9\") pod \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\" (UID: \"8bb988e1-ef1a-45b8-b3fc-cf6025835a89\") " Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.356257 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-kube-api-access-pmbq9" (OuterVolumeSpecName: "kube-api-access-pmbq9") pod "8bb988e1-ef1a-45b8-b3fc-cf6025835a89" (UID: "8bb988e1-ef1a-45b8-b3fc-cf6025835a89"). InnerVolumeSpecName "kube-api-access-pmbq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.381865 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bb988e1-ef1a-45b8-b3fc-cf6025835a89" (UID: "8bb988e1-ef1a-45b8-b3fc-cf6025835a89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.398592 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-config-data" (OuterVolumeSpecName: "config-data") pod "8bb988e1-ef1a-45b8-b3fc-cf6025835a89" (UID: "8bb988e1-ef1a-45b8-b3fc-cf6025835a89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.450913 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.451178 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.451642 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmbq9\" (UniqueName: \"kubernetes.io/projected/8bb988e1-ef1a-45b8-b3fc-cf6025835a89-kube-api-access-pmbq9\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.567886 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.631660 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.647232 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.658788 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:38:20 crc kubenswrapper[4744]: E1201 08:38:20.659221 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb988e1-ef1a-45b8-b3fc-cf6025835a89" containerName="nova-scheduler-scheduler" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.659239 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb988e1-ef1a-45b8-b3fc-cf6025835a89" containerName="nova-scheduler-scheduler" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.659438 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb988e1-ef1a-45b8-b3fc-cf6025835a89" containerName="nova-scheduler-scheduler" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.660060 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.662235 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.671270 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.756071 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfxdl\" (UniqueName: \"kubernetes.io/projected/97f2638c-75e4-4b25-be41-16e6b261c1b3-kube-api-access-cfxdl\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.756294 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97f2638c-75e4-4b25-be41-16e6b261c1b3-config-data\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.756358 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97f2638c-75e4-4b25-be41-16e6b261c1b3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.859160 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfxdl\" (UniqueName: \"kubernetes.io/projected/97f2638c-75e4-4b25-be41-16e6b261c1b3-kube-api-access-cfxdl\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.859215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97f2638c-75e4-4b25-be41-16e6b261c1b3-config-data\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.859306 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97f2638c-75e4-4b25-be41-16e6b261c1b3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.864367 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97f2638c-75e4-4b25-be41-16e6b261c1b3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.865286 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97f2638c-75e4-4b25-be41-16e6b261c1b3-config-data\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.903365 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfxdl\" (UniqueName: \"kubernetes.io/projected/97f2638c-75e4-4b25-be41-16e6b261c1b3-kube-api-access-cfxdl\") pod \"nova-scheduler-0\" (UID: \"97f2638c-75e4-4b25-be41-16e6b261c1b3\") " pod="openstack/nova-scheduler-0" Dec 01 08:38:20 crc kubenswrapper[4744]: I1201 08:38:20.969306 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 08:38:21 crc kubenswrapper[4744]: I1201 08:38:21.313772 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec302fdc-d505-4f9c-b403-14643c501988","Type":"ContainerStarted","Data":"7a9945b14742d4e870ad379c4d8440fd78be42ce2de8b9d2d731c1438909a6fb"} Dec 01 08:38:21 crc kubenswrapper[4744]: I1201 08:38:21.314111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec302fdc-d505-4f9c-b403-14643c501988","Type":"ContainerStarted","Data":"453f5eb152421af6405b00575bc9c76e00c076c75cb08676006ad097c597b9f4"} Dec 01 08:38:21 crc kubenswrapper[4744]: I1201 08:38:21.314128 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec302fdc-d505-4f9c-b403-14643c501988","Type":"ContainerStarted","Data":"eeb4d9aebca751e163b2e3163e4faad076feb9ec0ffbc3f21994ba04468499e8"} Dec 01 08:38:21 crc kubenswrapper[4744]: I1201 08:38:21.337614 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.337596511 podStartE2EDuration="2.337596511s" podCreationTimestamp="2025-12-01 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:38:21.331505627 +0000 UTC m=+1233.320563548" watchObservedRunningTime="2025-12-01 08:38:21.337596511 +0000 UTC m=+1233.326654432" Dec 01 08:38:21 crc kubenswrapper[4744]: I1201 08:38:21.448480 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 08:38:21 crc kubenswrapper[4744]: W1201 08:38:21.450047 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97f2638c_75e4_4b25_be41_16e6b261c1b3.slice/crio-536ac8c9066a1effcffa464e125595e942aa92bc695d8e75958b9812d44e60d8 WatchSource:0}: Error finding container 536ac8c9066a1effcffa464e125595e942aa92bc695d8e75958b9812d44e60d8: Status 404 returned error can't find the container with id 536ac8c9066a1effcffa464e125595e942aa92bc695d8e75958b9812d44e60d8 Dec 01 08:38:21 crc kubenswrapper[4744]: I1201 08:38:21.657258 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:56862->10.217.0.202:8775: read: connection reset by peer" Dec 01 08:38:21 crc kubenswrapper[4744]: I1201 08:38:21.658217 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:56866->10.217.0.202:8775: read: connection reset by peer" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.182898 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.291255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-combined-ca-bundle\") pod \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.291469 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-config-data\") pod \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.291598 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-nova-metadata-tls-certs\") pod \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.291679 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prv6c\" (UniqueName: \"kubernetes.io/projected/19dbb279-f666-4a27-a699-88ccd0b5cf2b-kube-api-access-prv6c\") pod \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.291724 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19dbb279-f666-4a27-a699-88ccd0b5cf2b-logs\") pod \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\" (UID: \"19dbb279-f666-4a27-a699-88ccd0b5cf2b\") " Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.292271 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19dbb279-f666-4a27-a699-88ccd0b5cf2b-logs" (OuterVolumeSpecName: "logs") pod "19dbb279-f666-4a27-a699-88ccd0b5cf2b" (UID: "19dbb279-f666-4a27-a699-88ccd0b5cf2b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.306870 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bb988e1-ef1a-45b8-b3fc-cf6025835a89" path="/var/lib/kubelet/pods/8bb988e1-ef1a-45b8-b3fc-cf6025835a89/volumes" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.311821 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19dbb279-f666-4a27-a699-88ccd0b5cf2b-kube-api-access-prv6c" (OuterVolumeSpecName: "kube-api-access-prv6c") pod "19dbb279-f666-4a27-a699-88ccd0b5cf2b" (UID: "19dbb279-f666-4a27-a699-88ccd0b5cf2b"). InnerVolumeSpecName "kube-api-access-prv6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.327011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-config-data" (OuterVolumeSpecName: "config-data") pod "19dbb279-f666-4a27-a699-88ccd0b5cf2b" (UID: "19dbb279-f666-4a27-a699-88ccd0b5cf2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.329314 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19dbb279-f666-4a27-a699-88ccd0b5cf2b" (UID: "19dbb279-f666-4a27-a699-88ccd0b5cf2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.332264 4744 generic.go:334] "Generic (PLEG): container finished" podID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerID="9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd" exitCode=0 Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.332884 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.352458 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.352440924 podStartE2EDuration="2.352440924s" podCreationTimestamp="2025-12-01 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:38:22.348896114 +0000 UTC m=+1234.337954045" watchObservedRunningTime="2025-12-01 08:38:22.352440924 +0000 UTC m=+1234.341498845" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.357544 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "19dbb279-f666-4a27-a699-88ccd0b5cf2b" (UID: "19dbb279-f666-4a27-a699-88ccd0b5cf2b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.394019 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.394217 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.394278 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prv6c\" (UniqueName: \"kubernetes.io/projected/19dbb279-f666-4a27-a699-88ccd0b5cf2b-kube-api-access-prv6c\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.394356 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19dbb279-f666-4a27-a699-88ccd0b5cf2b-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.394430 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dbb279-f666-4a27-a699-88ccd0b5cf2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.421113 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"97f2638c-75e4-4b25-be41-16e6b261c1b3","Type":"ContainerStarted","Data":"894016b091d9069fd28541cf745e069d5497a9c3fe3134319191585a3f7f13c8"} Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.421351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"97f2638c-75e4-4b25-be41-16e6b261c1b3","Type":"ContainerStarted","Data":"536ac8c9066a1effcffa464e125595e942aa92bc695d8e75958b9812d44e60d8"} Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.421494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19dbb279-f666-4a27-a699-88ccd0b5cf2b","Type":"ContainerDied","Data":"9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd"} Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.421583 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19dbb279-f666-4a27-a699-88ccd0b5cf2b","Type":"ContainerDied","Data":"8425bb9558f15b8cb90598cca9e4767f9d6bc701a8623b15119b6bce88cc77c1"} Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.421692 4744 scope.go:117] "RemoveContainer" containerID="9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.487567 4744 scope.go:117] "RemoveContainer" containerID="b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.528464 4744 scope.go:117] "RemoveContainer" containerID="9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd" Dec 01 08:38:22 crc kubenswrapper[4744]: E1201 08:38:22.528957 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd\": container with ID starting with 9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd not found: ID does not exist" containerID="9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.529047 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd"} err="failed to get container status \"9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd\": rpc error: code = NotFound desc = could not find container \"9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd\": container with ID starting with 9112372be320fe4024c24d27b986bfa8c0ae682a14dbfa93731f6a1a860dd3fd not found: ID does not exist" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.529135 4744 scope.go:117] "RemoveContainer" containerID="b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059" Dec 01 08:38:22 crc kubenswrapper[4744]: E1201 08:38:22.529577 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059\": container with ID starting with b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059 not found: ID does not exist" containerID="b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.529634 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059"} err="failed to get container status \"b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059\": rpc error: code = NotFound desc = could not find container \"b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059\": container with ID starting with b277d987c7c7424fa11238751bf290fffc63be23b06775f94344f95085372059 not found: ID does not exist" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.667605 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.677308 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.693803 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:38:22 crc kubenswrapper[4744]: E1201 08:38:22.694324 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-metadata" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.694351 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-metadata" Dec 01 08:38:22 crc kubenswrapper[4744]: E1201 08:38:22.694363 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-log" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.694372 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-log" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.695298 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-log" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.695341 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" containerName="nova-metadata-metadata" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.696466 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.702526 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.702833 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.708547 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.801894 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.801946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-logs\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.802081 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-config-data\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.802186 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.802244 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjlsf\" (UniqueName: \"kubernetes.io/projected/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-kube-api-access-cjlsf\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.904104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.904180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjlsf\" (UniqueName: \"kubernetes.io/projected/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-kube-api-access-cjlsf\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.904213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.904241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-logs\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.904316 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-config-data\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.905514 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-logs\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.908115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.908604 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-config-data\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.910720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:22 crc kubenswrapper[4744]: I1201 08:38:22.931029 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjlsf\" (UniqueName: \"kubernetes.io/projected/7bfdbce6-2016-4da0-8d18-300bbc3e92fb-kube-api-access-cjlsf\") pod \"nova-metadata-0\" (UID: \"7bfdbce6-2016-4da0-8d18-300bbc3e92fb\") " pod="openstack/nova-metadata-0" Dec 01 08:38:23 crc kubenswrapper[4744]: I1201 08:38:23.019315 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 08:38:23 crc kubenswrapper[4744]: I1201 08:38:23.468808 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 08:38:23 crc kubenswrapper[4744]: W1201 08:38:23.472061 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bfdbce6_2016_4da0_8d18_300bbc3e92fb.slice/crio-428060772baabbce4de80f0ce1ab7e0c16ca5fe6bde4b898372819e887bf5052 WatchSource:0}: Error finding container 428060772baabbce4de80f0ce1ab7e0c16ca5fe6bde4b898372819e887bf5052: Status 404 returned error can't find the container with id 428060772baabbce4de80f0ce1ab7e0c16ca5fe6bde4b898372819e887bf5052 Dec 01 08:38:24 crc kubenswrapper[4744]: I1201 08:38:24.325697 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19dbb279-f666-4a27-a699-88ccd0b5cf2b" path="/var/lib/kubelet/pods/19dbb279-f666-4a27-a699-88ccd0b5cf2b/volumes" Dec 01 08:38:24 crc kubenswrapper[4744]: I1201 08:38:24.353358 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7bfdbce6-2016-4da0-8d18-300bbc3e92fb","Type":"ContainerStarted","Data":"204993ca96186c307e3adbd84ec5ca0e055ec8bdbed678391372b38b1faf49cc"} Dec 01 08:38:24 crc kubenswrapper[4744]: I1201 08:38:24.353520 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7bfdbce6-2016-4da0-8d18-300bbc3e92fb","Type":"ContainerStarted","Data":"480234d46fdfcd530c8d6a048b090c4cad70758509f30d8990b3ecf9b253c355"} Dec 01 08:38:24 crc kubenswrapper[4744]: I1201 08:38:24.353535 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7bfdbce6-2016-4da0-8d18-300bbc3e92fb","Type":"ContainerStarted","Data":"428060772baabbce4de80f0ce1ab7e0c16ca5fe6bde4b898372819e887bf5052"} Dec 01 08:38:24 crc kubenswrapper[4744]: I1201 08:38:24.379620 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.379583751 podStartE2EDuration="2.379583751s" podCreationTimestamp="2025-12-01 08:38:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:38:24.372622992 +0000 UTC m=+1236.361680953" watchObservedRunningTime="2025-12-01 08:38:24.379583751 +0000 UTC m=+1236.368641672" Dec 01 08:38:25 crc kubenswrapper[4744]: I1201 08:38:25.970187 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 08:38:28 crc kubenswrapper[4744]: I1201 08:38:28.020340 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 08:38:28 crc kubenswrapper[4744]: I1201 08:38:28.021107 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 08:38:30 crc kubenswrapper[4744]: I1201 08:38:30.061646 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 08:38:30 crc kubenswrapper[4744]: I1201 08:38:30.062493 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 08:38:30 crc kubenswrapper[4744]: I1201 08:38:30.970432 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 08:38:31 crc kubenswrapper[4744]: I1201 08:38:31.033259 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 08:38:31 crc kubenswrapper[4744]: I1201 08:38:31.075599 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ec302fdc-d505-4f9c-b403-14643c501988" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 08:38:31 crc kubenswrapper[4744]: I1201 08:38:31.075681 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ec302fdc-d505-4f9c-b403-14643c501988" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 08:38:31 crc kubenswrapper[4744]: I1201 08:38:31.474914 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 08:38:33 crc kubenswrapper[4744]: I1201 08:38:33.020112 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 08:38:33 crc kubenswrapper[4744]: I1201 08:38:33.020188 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 08:38:34 crc kubenswrapper[4744]: I1201 08:38:34.032797 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7bfdbce6-2016-4da0-8d18-300bbc3e92fb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 08:38:34 crc kubenswrapper[4744]: I1201 08:38:34.032835 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7bfdbce6-2016-4da0-8d18-300bbc3e92fb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 08:38:37 crc kubenswrapper[4744]: I1201 08:38:37.575017 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 08:38:40 crc kubenswrapper[4744]: I1201 08:38:40.069574 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 08:38:40 crc kubenswrapper[4744]: I1201 08:38:40.070264 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 08:38:40 crc kubenswrapper[4744]: I1201 08:38:40.071513 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 08:38:40 crc kubenswrapper[4744]: I1201 08:38:40.079790 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 08:38:40 crc kubenswrapper[4744]: I1201 08:38:40.575019 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 08:38:40 crc kubenswrapper[4744]: I1201 08:38:40.597137 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 08:38:43 crc kubenswrapper[4744]: I1201 08:38:43.025285 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 08:38:43 crc kubenswrapper[4744]: I1201 08:38:43.027589 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 08:38:43 crc kubenswrapper[4744]: I1201 08:38:43.035113 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 08:38:43 crc kubenswrapper[4744]: I1201 08:38:43.614682 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 08:38:51 crc kubenswrapper[4744]: I1201 08:38:51.992152 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:38:53 crc kubenswrapper[4744]: I1201 08:38:53.709127 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:38:55 crc kubenswrapper[4744]: I1201 08:38:55.951060 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerName="rabbitmq" containerID="cri-o://2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993" gracePeriod=604797 Dec 01 08:38:57 crc kubenswrapper[4744]: I1201 08:38:57.676502 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="788a0c2b-cd13-48cb-99af-ba838e154525" containerName="rabbitmq" containerID="cri-o://55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3" gracePeriod=604797 Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.574012 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.674894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-confd\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.674959 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h58q\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-kube-api-access-5h58q\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675029 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-erlang-cookie\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675124 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675156 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf94bcff-f22a-4a35-beca-0096f08ee1f6-erlang-cookie-secret\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675188 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-config-data\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675248 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf94bcff-f22a-4a35-beca-0096f08ee1f6-pod-info\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675294 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-plugins\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675366 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-server-conf\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675390 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-plugins-conf\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.675468 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-tls\") pod \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\" (UID: \"bf94bcff-f22a-4a35-beca-0096f08ee1f6\") " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.676354 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.679321 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.682691 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.684261 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.686739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-kube-api-access-5h58q" (OuterVolumeSpecName: "kube-api-access-5h58q") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "kube-api-access-5h58q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.688450 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.691587 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf94bcff-f22a-4a35-beca-0096f08ee1f6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.713027 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bf94bcff-f22a-4a35-beca-0096f08ee1f6-pod-info" (OuterVolumeSpecName: "pod-info") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.727450 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-config-data" (OuterVolumeSpecName: "config-data") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778052 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h58q\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-kube-api-access-5h58q\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778088 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778111 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778122 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf94bcff-f22a-4a35-beca-0096f08ee1f6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778131 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778139 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf94bcff-f22a-4a35-beca-0096f08ee1f6-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778147 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778155 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.778162 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.797737 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-server-conf" (OuterVolumeSpecName: "server-conf") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.805582 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.812690 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerID="2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993" exitCode=0 Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.812739 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bf94bcff-f22a-4a35-beca-0096f08ee1f6","Type":"ContainerDied","Data":"2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993"} Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.812774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bf94bcff-f22a-4a35-beca-0096f08ee1f6","Type":"ContainerDied","Data":"f60789b4e08fa98fbc562d71cfeeefd3c88476b25244c6a2ed8278cbea2a0b9c"} Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.812794 4744 scope.go:117] "RemoveContainer" containerID="2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.812943 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.830783 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "bf94bcff-f22a-4a35-beca-0096f08ee1f6" (UID: "bf94bcff-f22a-4a35-beca-0096f08ee1f6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.846164 4744 scope.go:117] "RemoveContainer" containerID="cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.874742 4744 scope.go:117] "RemoveContainer" containerID="2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993" Dec 01 08:39:02 crc kubenswrapper[4744]: E1201 08:39:02.875207 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993\": container with ID starting with 2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993 not found: ID does not exist" containerID="2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.875252 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993"} err="failed to get container status \"2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993\": rpc error: code = NotFound desc = could not find container \"2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993\": container with ID starting with 2f4436d1e08994a7fec2f5b1bfbd75103386b3057d1532b330585ab422b58993 not found: ID does not exist" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.875280 4744 scope.go:117] "RemoveContainer" containerID="cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d" Dec 01 08:39:02 crc kubenswrapper[4744]: E1201 08:39:02.875611 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d\": container with ID starting with cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d not found: ID does not exist" containerID="cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.875663 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d"} err="failed to get container status \"cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d\": rpc error: code = NotFound desc = could not find container \"cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d\": container with ID starting with cba271c2ba161e2954c68fd169fe51496f4c02fa1923189e4d99b9f41db9529d not found: ID does not exist" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.879822 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf94bcff-f22a-4a35-beca-0096f08ee1f6-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.879844 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf94bcff-f22a-4a35-beca-0096f08ee1f6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:02 crc kubenswrapper[4744]: I1201 08:39:02.879855 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.147727 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.162279 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.173842 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:39:03 crc kubenswrapper[4744]: E1201 08:39:03.174188 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerName="rabbitmq" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.174204 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerName="rabbitmq" Dec 01 08:39:03 crc kubenswrapper[4744]: E1201 08:39:03.174231 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerName="setup-container" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.174237 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerName="setup-container" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.174493 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" containerName="rabbitmq" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.175390 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.178259 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.178294 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.178508 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.178595 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.178927 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x8gr9" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.178858 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.182183 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.195121 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288284 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288331 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93e40ea3-265b-44a9-ada5-5b583201148d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288365 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc4db\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-kube-api-access-fc4db\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288499 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-config-data\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288521 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288638 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93e40ea3-265b-44a9-ada5-5b583201148d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288784 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288838 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.288992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391166 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391274 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93e40ea3-265b-44a9-ada5-5b583201148d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc4db\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-kube-api-access-fc4db\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391355 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-config-data\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391376 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391401 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391434 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93e40ea3-265b-44a9-ada5-5b583201148d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391491 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.391713 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.393238 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.396042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.397074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.397319 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.397432 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.397662 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e40ea3-265b-44a9-ada5-5b583201148d-config-data\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.400159 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.400241 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93e40ea3-265b-44a9-ada5-5b583201148d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.401201 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93e40ea3-265b-44a9-ada5-5b583201148d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.412956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc4db\" (UniqueName: \"kubernetes.io/projected/93e40ea3-265b-44a9-ada5-5b583201148d-kube-api-access-fc4db\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.434255 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"93e40ea3-265b-44a9-ada5-5b583201148d\") " pod="openstack/rabbitmq-server-0" Dec 01 08:39:03 crc kubenswrapper[4744]: I1201 08:39:03.546366 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.299617 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf94bcff-f22a-4a35-beca-0096f08ee1f6" path="/var/lib/kubelet/pods/bf94bcff-f22a-4a35-beca-0096f08ee1f6/volumes" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.593555 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.625427 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.727757 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-plugins-conf\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.727838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtkgl\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-kube-api-access-dtkgl\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.727892 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-plugins\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.727961 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.728007 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-erlang-cookie\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.728082 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-config-data\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.728147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-confd\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.728177 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/788a0c2b-cd13-48cb-99af-ba838e154525-erlang-cookie-secret\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.728205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-tls\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.728262 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/788a0c2b-cd13-48cb-99af-ba838e154525-pod-info\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.728301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-server-conf\") pod \"788a0c2b-cd13-48cb-99af-ba838e154525\" (UID: \"788a0c2b-cd13-48cb-99af-ba838e154525\") " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.729635 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.729681 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.730233 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.733535 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.733799 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-kube-api-access-dtkgl" (OuterVolumeSpecName: "kube-api-access-dtkgl") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "kube-api-access-dtkgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.733856 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.736247 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/788a0c2b-cd13-48cb-99af-ba838e154525-pod-info" (OuterVolumeSpecName: "pod-info") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.737846 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788a0c2b-cd13-48cb-99af-ba838e154525-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.753116 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-config-data" (OuterVolumeSpecName: "config-data") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.824460 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-server-conf" (OuterVolumeSpecName: "server-conf") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.829896 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830004 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/788a0c2b-cd13-48cb-99af-ba838e154525-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830089 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830169 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/788a0c2b-cd13-48cb-99af-ba838e154525-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830221 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830311 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/788a0c2b-cd13-48cb-99af-ba838e154525-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830373 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtkgl\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-kube-api-access-dtkgl\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830445 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830526 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.830578 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.837570 4744 generic.go:334] "Generic (PLEG): container finished" podID="788a0c2b-cd13-48cb-99af-ba838e154525" containerID="55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3" exitCode=0 Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.837645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"788a0c2b-cd13-48cb-99af-ba838e154525","Type":"ContainerDied","Data":"55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3"} Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.837674 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"788a0c2b-cd13-48cb-99af-ba838e154525","Type":"ContainerDied","Data":"7cff00ef7426c20912aca14f985a667dd253ac736be714672885555d071f5104"} Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.837691 4744 scope.go:117] "RemoveContainer" containerID="55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.837719 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.840170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93e40ea3-265b-44a9-ada5-5b583201148d","Type":"ContainerStarted","Data":"068ea29a4dde576b3654c4384498efb142efc478bb6e7b36a12c669b1b078575"} Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.851115 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "788a0c2b-cd13-48cb-99af-ba838e154525" (UID: "788a0c2b-cd13-48cb-99af-ba838e154525"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.869455 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.870399 4744 scope.go:117] "RemoveContainer" containerID="3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.893254 4744 scope.go:117] "RemoveContainer" containerID="55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3" Dec 01 08:39:04 crc kubenswrapper[4744]: E1201 08:39:04.893671 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3\": container with ID starting with 55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3 not found: ID does not exist" containerID="55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.893712 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3"} err="failed to get container status \"55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3\": rpc error: code = NotFound desc = could not find container \"55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3\": container with ID starting with 55a70003d64866c14c1f2f5236af2f7b36366a2b667bf39a5aa392502156b5c3 not found: ID does not exist" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.893740 4744 scope.go:117] "RemoveContainer" containerID="3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f" Dec 01 08:39:04 crc kubenswrapper[4744]: E1201 08:39:04.894011 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f\": container with ID starting with 3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f not found: ID does not exist" containerID="3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.894059 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f"} err="failed to get container status \"3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f\": rpc error: code = NotFound desc = could not find container \"3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f\": container with ID starting with 3ee206bc2ada9ac77b16351df4e51ee01f418bb8d1e7e7a88f936c6c89db988f not found: ID does not exist" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.932742 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:04 crc kubenswrapper[4744]: I1201 08:39:04.932778 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/788a0c2b-cd13-48cb-99af-ba838e154525-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.202552 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.220625 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.240258 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:39:05 crc kubenswrapper[4744]: E1201 08:39:05.240775 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a0c2b-cd13-48cb-99af-ba838e154525" containerName="rabbitmq" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.240796 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a0c2b-cd13-48cb-99af-ba838e154525" containerName="rabbitmq" Dec 01 08:39:05 crc kubenswrapper[4744]: E1201 08:39:05.240834 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a0c2b-cd13-48cb-99af-ba838e154525" containerName="setup-container" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.240843 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a0c2b-cd13-48cb-99af-ba838e154525" containerName="setup-container" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.241071 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="788a0c2b-cd13-48cb-99af-ba838e154525" containerName="rabbitmq" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.242297 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.244879 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.245042 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.245834 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.245940 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.246050 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.246324 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.246699 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-75bjt" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.259764 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.342833 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g86n7\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-kube-api-access-g86n7\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.342894 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a93bac-aabe-4109-950f-14cb6ac634dd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.342918 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a93bac-aabe-4109-950f-14cb6ac634dd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343112 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343136 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343167 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343277 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.343348 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.444768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445066 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g86n7\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-kube-api-access-g86n7\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445164 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a93bac-aabe-4109-950f-14cb6ac634dd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a93bac-aabe-4109-950f-14cb6ac634dd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445288 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445363 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.445608 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.446232 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.446252 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.446872 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.449758 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.450151 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a93bac-aabe-4109-950f-14cb6ac634dd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.457591 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a93bac-aabe-4109-950f-14cb6ac634dd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.457764 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.458148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.458934 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a93bac-aabe-4109-950f-14cb6ac634dd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.463294 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g86n7\" (UniqueName: \"kubernetes.io/projected/e4a93bac-aabe-4109-950f-14cb6ac634dd-kube-api-access-g86n7\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.486503 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e4a93bac-aabe-4109-950f-14cb6ac634dd\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:05 crc kubenswrapper[4744]: I1201 08:39:05.587558 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.020078 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 08:39:06 crc kubenswrapper[4744]: W1201 08:39:06.027283 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4a93bac_aabe_4109_950f_14cb6ac634dd.slice/crio-776a9bd4b914da7f5e673e4d151510132a9b263c92fb87a49a194da530fe66f0 WatchSource:0}: Error finding container 776a9bd4b914da7f5e673e4d151510132a9b263c92fb87a49a194da530fe66f0: Status 404 returned error can't find the container with id 776a9bd4b914da7f5e673e4d151510132a9b263c92fb87a49a194da530fe66f0 Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.307299 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788a0c2b-cd13-48cb-99af-ba838e154525" path="/var/lib/kubelet/pods/788a0c2b-cd13-48cb-99af-ba838e154525/volumes" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.465882 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-x5f4g"] Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.467825 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.476894 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.478687 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-x5f4g"] Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.565302 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x729g\" (UniqueName: \"kubernetes.io/projected/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-kube-api-access-x729g\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.565474 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.565562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.565694 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-config\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.565766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.565992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.566059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-svc\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.668017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x729g\" (UniqueName: \"kubernetes.io/projected/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-kube-api-access-x729g\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.668088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.668146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.668215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-config\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.668268 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.668470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.668525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-svc\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.669196 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.669198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.669556 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-svc\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.669815 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.669938 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-config\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.670449 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.702244 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x729g\" (UniqueName: \"kubernetes.io/projected/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-kube-api-access-x729g\") pod \"dnsmasq-dns-594cb89c79-x5f4g\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.788340 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.869276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93e40ea3-265b-44a9-ada5-5b583201148d","Type":"ContainerStarted","Data":"ad795d047798f0653d4837b85e3555080dea221f31ae80a1b9683947819a30d7"} Dec 01 08:39:06 crc kubenswrapper[4744]: I1201 08:39:06.872083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4a93bac-aabe-4109-950f-14cb6ac634dd","Type":"ContainerStarted","Data":"776a9bd4b914da7f5e673e4d151510132a9b263c92fb87a49a194da530fe66f0"} Dec 01 08:39:07 crc kubenswrapper[4744]: I1201 08:39:07.301026 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-x5f4g"] Dec 01 08:39:07 crc kubenswrapper[4744]: W1201 08:39:07.381225 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e6ef6aa_a18c_4dac_81c8_736c37f0e203.slice/crio-e91341fdfba27f3ea788d806a6473aba61e2eb29c2155d741b5fa7c83f431358 WatchSource:0}: Error finding container e91341fdfba27f3ea788d806a6473aba61e2eb29c2155d741b5fa7c83f431358: Status 404 returned error can't find the container with id e91341fdfba27f3ea788d806a6473aba61e2eb29c2155d741b5fa7c83f431358 Dec 01 08:39:07 crc kubenswrapper[4744]: I1201 08:39:07.886561 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4a93bac-aabe-4109-950f-14cb6ac634dd","Type":"ContainerStarted","Data":"6230bf20d2289d169a27b1d6f81f3bf894ede30d1338c678de3ae2da07101643"} Dec 01 08:39:07 crc kubenswrapper[4744]: I1201 08:39:07.889549 4744 generic.go:334] "Generic (PLEG): container finished" podID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerID="1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c" exitCode=0 Dec 01 08:39:07 crc kubenswrapper[4744]: I1201 08:39:07.889695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" event={"ID":"5e6ef6aa-a18c-4dac-81c8-736c37f0e203","Type":"ContainerDied","Data":"1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c"} Dec 01 08:39:07 crc kubenswrapper[4744]: I1201 08:39:07.889786 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" event={"ID":"5e6ef6aa-a18c-4dac-81c8-736c37f0e203","Type":"ContainerStarted","Data":"e91341fdfba27f3ea788d806a6473aba61e2eb29c2155d741b5fa7c83f431358"} Dec 01 08:39:08 crc kubenswrapper[4744]: I1201 08:39:08.902499 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" event={"ID":"5e6ef6aa-a18c-4dac-81c8-736c37f0e203","Type":"ContainerStarted","Data":"be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b"} Dec 01 08:39:08 crc kubenswrapper[4744]: I1201 08:39:08.902897 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:08 crc kubenswrapper[4744]: I1201 08:39:08.950062 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" podStartSLOduration=2.950036948 podStartE2EDuration="2.950036948s" podCreationTimestamp="2025-12-01 08:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:39:08.931356125 +0000 UTC m=+1280.920414086" watchObservedRunningTime="2025-12-01 08:39:08.950036948 +0000 UTC m=+1280.939094909" Dec 01 08:39:16 crc kubenswrapper[4744]: I1201 08:39:16.790689 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:16 crc kubenswrapper[4744]: I1201 08:39:16.878219 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-5q242"] Dec 01 08:39:16 crc kubenswrapper[4744]: I1201 08:39:16.878526 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" podUID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerName="dnsmasq-dns" containerID="cri-o://e344853386235a5230a3c33ae9ce75dbe0ee43c71013f2662845ef0a9b12b195" gracePeriod=10 Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.008377 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-5nlp9"] Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.010180 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.015927 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerID="e344853386235a5230a3c33ae9ce75dbe0ee43c71013f2662845ef0a9b12b195" exitCode=0 Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.015979 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" event={"ID":"2b4ff7ad-86db-42be-bb0f-9347c1c0494c","Type":"ContainerDied","Data":"e344853386235a5230a3c33ae9ce75dbe0ee43c71013f2662845ef0a9b12b195"} Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.028928 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-5nlp9"] Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.129315 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.129429 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.129472 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.129549 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-config\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.129629 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.129678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwtng\" (UniqueName: \"kubernetes.io/projected/dee5122c-f685-44a1-acd0-9fcfe3968bb7-kube-api-access-nwtng\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.129887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.231743 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232039 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232110 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232148 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-config\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232201 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwtng\" (UniqueName: \"kubernetes.io/projected/dee5122c-f685-44a1-acd0-9fcfe3968bb7-kube-api-access-nwtng\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232565 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.232975 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.233760 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-config\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.234000 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.234083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.235790 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dee5122c-f685-44a1-acd0-9fcfe3968bb7-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.254908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwtng\" (UniqueName: \"kubernetes.io/projected/dee5122c-f685-44a1-acd0-9fcfe3968bb7-kube-api-access-nwtng\") pod \"dnsmasq-dns-5596c69fcc-5nlp9\" (UID: \"dee5122c-f685-44a1-acd0-9fcfe3968bb7\") " pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.368796 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.378181 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.537837 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhtkx\" (UniqueName: \"kubernetes.io/projected/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-kube-api-access-jhtkx\") pod \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.538180 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-svc\") pod \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.538234 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-swift-storage-0\") pod \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.538260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-sb\") pod \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.538280 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-nb\") pod \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.538303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-config\") pod \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\" (UID: \"2b4ff7ad-86db-42be-bb0f-9347c1c0494c\") " Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.545349 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-kube-api-access-jhtkx" (OuterVolumeSpecName: "kube-api-access-jhtkx") pod "2b4ff7ad-86db-42be-bb0f-9347c1c0494c" (UID: "2b4ff7ad-86db-42be-bb0f-9347c1c0494c"). InnerVolumeSpecName "kube-api-access-jhtkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.590151 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2b4ff7ad-86db-42be-bb0f-9347c1c0494c" (UID: "2b4ff7ad-86db-42be-bb0f-9347c1c0494c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.594536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2b4ff7ad-86db-42be-bb0f-9347c1c0494c" (UID: "2b4ff7ad-86db-42be-bb0f-9347c1c0494c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.599452 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2b4ff7ad-86db-42be-bb0f-9347c1c0494c" (UID: "2b4ff7ad-86db-42be-bb0f-9347c1c0494c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.609718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b4ff7ad-86db-42be-bb0f-9347c1c0494c" (UID: "2b4ff7ad-86db-42be-bb0f-9347c1c0494c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.624128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-config" (OuterVolumeSpecName: "config") pod "2b4ff7ad-86db-42be-bb0f-9347c1c0494c" (UID: "2b4ff7ad-86db-42be-bb0f-9347c1c0494c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.641039 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.641070 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.641086 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.641097 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.641108 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.641119 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhtkx\" (UniqueName: \"kubernetes.io/projected/2b4ff7ad-86db-42be-bb0f-9347c1c0494c-kube-api-access-jhtkx\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:17 crc kubenswrapper[4744]: I1201 08:39:17.853864 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-5nlp9"] Dec 01 08:39:17 crc kubenswrapper[4744]: W1201 08:39:17.854958 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddee5122c_f685_44a1_acd0_9fcfe3968bb7.slice/crio-3ee53d70fe1c5955e94f1381584cb99da00e287fe1166814f55dfdd85895f5e0 WatchSource:0}: Error finding container 3ee53d70fe1c5955e94f1381584cb99da00e287fe1166814f55dfdd85895f5e0: Status 404 returned error can't find the container with id 3ee53d70fe1c5955e94f1381584cb99da00e287fe1166814f55dfdd85895f5e0 Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.026441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" event={"ID":"dee5122c-f685-44a1-acd0-9fcfe3968bb7","Type":"ContainerStarted","Data":"3ee53d70fe1c5955e94f1381584cb99da00e287fe1166814f55dfdd85895f5e0"} Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.028803 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.028806 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-5q242" event={"ID":"2b4ff7ad-86db-42be-bb0f-9347c1c0494c","Type":"ContainerDied","Data":"b1b0f3b0684c14178aacfb6c5ecbab017952ebf6b705bb6e382e4d97ec41266d"} Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.028848 4744 scope.go:117] "RemoveContainer" containerID="e344853386235a5230a3c33ae9ce75dbe0ee43c71013f2662845ef0a9b12b195" Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.154553 4744 scope.go:117] "RemoveContainer" containerID="6f4ccadcd61c8ca2195b3c260389ade4801be0f481c699f658e28fe7d986cec2" Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.161205 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-5q242"] Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.171867 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-5q242"] Dec 01 08:39:18 crc kubenswrapper[4744]: I1201 08:39:18.300498 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" path="/var/lib/kubelet/pods/2b4ff7ad-86db-42be-bb0f-9347c1c0494c/volumes" Dec 01 08:39:19 crc kubenswrapper[4744]: I1201 08:39:19.076228 4744 generic.go:334] "Generic (PLEG): container finished" podID="dee5122c-f685-44a1-acd0-9fcfe3968bb7" containerID="7b137ec508f351a69accbefd9f813ca0a898239b121eab3674ca2a96475829c9" exitCode=0 Dec 01 08:39:19 crc kubenswrapper[4744]: I1201 08:39:19.076283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" event={"ID":"dee5122c-f685-44a1-acd0-9fcfe3968bb7","Type":"ContainerDied","Data":"7b137ec508f351a69accbefd9f813ca0a898239b121eab3674ca2a96475829c9"} Dec 01 08:39:20 crc kubenswrapper[4744]: I1201 08:39:20.092100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" event={"ID":"dee5122c-f685-44a1-acd0-9fcfe3968bb7","Type":"ContainerStarted","Data":"06ffbb8d6d799073c6a931e661b738dbe45a40cd584c34a770d8eb9ee65e4dde"} Dec 01 08:39:20 crc kubenswrapper[4744]: I1201 08:39:20.092373 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:20 crc kubenswrapper[4744]: I1201 08:39:20.114603 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" podStartSLOduration=4.114580118 podStartE2EDuration="4.114580118s" podCreationTimestamp="2025-12-01 08:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:39:20.111442158 +0000 UTC m=+1292.100500109" watchObservedRunningTime="2025-12-01 08:39:20.114580118 +0000 UTC m=+1292.103638039" Dec 01 08:39:27 crc kubenswrapper[4744]: I1201 08:39:27.380390 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5596c69fcc-5nlp9" Dec 01 08:39:27 crc kubenswrapper[4744]: I1201 08:39:27.455306 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-x5f4g"] Dec 01 08:39:27 crc kubenswrapper[4744]: I1201 08:39:27.455562 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" podUID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerName="dnsmasq-dns" containerID="cri-o://be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b" gracePeriod=10 Dec 01 08:39:27 crc kubenswrapper[4744]: I1201 08:39:27.965649 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.063784 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-openstack-edpm-ipam\") pod \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.064155 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-nb\") pod \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.064184 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-sb\") pod \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.064251 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-swift-storage-0\") pod \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.064275 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x729g\" (UniqueName: \"kubernetes.io/projected/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-kube-api-access-x729g\") pod \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.064317 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-config\") pod \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.064349 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-svc\") pod \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\" (UID: \"5e6ef6aa-a18c-4dac-81c8-736c37f0e203\") " Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.074786 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-kube-api-access-x729g" (OuterVolumeSpecName: "kube-api-access-x729g") pod "5e6ef6aa-a18c-4dac-81c8-736c37f0e203" (UID: "5e6ef6aa-a18c-4dac-81c8-736c37f0e203"). InnerVolumeSpecName "kube-api-access-x729g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.122913 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5e6ef6aa-a18c-4dac-81c8-736c37f0e203" (UID: "5e6ef6aa-a18c-4dac-81c8-736c37f0e203"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.123251 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5e6ef6aa-a18c-4dac-81c8-736c37f0e203" (UID: "5e6ef6aa-a18c-4dac-81c8-736c37f0e203"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.123964 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5e6ef6aa-a18c-4dac-81c8-736c37f0e203" (UID: "5e6ef6aa-a18c-4dac-81c8-736c37f0e203"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.126159 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-config" (OuterVolumeSpecName: "config") pod "5e6ef6aa-a18c-4dac-81c8-736c37f0e203" (UID: "5e6ef6aa-a18c-4dac-81c8-736c37f0e203"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.130490 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5e6ef6aa-a18c-4dac-81c8-736c37f0e203" (UID: "5e6ef6aa-a18c-4dac-81c8-736c37f0e203"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.143672 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5e6ef6aa-a18c-4dac-81c8-736c37f0e203" (UID: "5e6ef6aa-a18c-4dac-81c8-736c37f0e203"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.166339 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.166394 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.166423 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.166436 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.166447 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x729g\" (UniqueName: \"kubernetes.io/projected/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-kube-api-access-x729g\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.166460 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.166470 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e6ef6aa-a18c-4dac-81c8-736c37f0e203-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.172378 4744 generic.go:334] "Generic (PLEG): container finished" podID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerID="be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b" exitCode=0 Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.172466 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.172454 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" event={"ID":"5e6ef6aa-a18c-4dac-81c8-736c37f0e203","Type":"ContainerDied","Data":"be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b"} Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.172603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-x5f4g" event={"ID":"5e6ef6aa-a18c-4dac-81c8-736c37f0e203","Type":"ContainerDied","Data":"e91341fdfba27f3ea788d806a6473aba61e2eb29c2155d741b5fa7c83f431358"} Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.172625 4744 scope.go:117] "RemoveContainer" containerID="be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.203535 4744 scope.go:117] "RemoveContainer" containerID="1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.216026 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-x5f4g"] Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.230507 4744 scope.go:117] "RemoveContainer" containerID="be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b" Dec 01 08:39:28 crc kubenswrapper[4744]: E1201 08:39:28.231363 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b\": container with ID starting with be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b not found: ID does not exist" containerID="be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.231419 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b"} err="failed to get container status \"be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b\": rpc error: code = NotFound desc = could not find container \"be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b\": container with ID starting with be65b7de0e5f53357bbb086a6fcae6f5f0feaee35374facb8e661688b743d04b not found: ID does not exist" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.231445 4744 scope.go:117] "RemoveContainer" containerID="1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c" Dec 01 08:39:28 crc kubenswrapper[4744]: E1201 08:39:28.232073 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c\": container with ID starting with 1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c not found: ID does not exist" containerID="1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.232121 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c"} err="failed to get container status \"1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c\": rpc error: code = NotFound desc = could not find container \"1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c\": container with ID starting with 1614c001b1d25d07cf9e1bdd2499f2059530de678729b281f1508958f0aba61c not found: ID does not exist" Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.233573 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-x5f4g"] Dec 01 08:39:28 crc kubenswrapper[4744]: I1201 08:39:28.297663 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" path="/var/lib/kubelet/pods/5e6ef6aa-a18c-4dac-81c8-736c37f0e203/volumes" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.143668 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz"] Dec 01 08:39:36 crc kubenswrapper[4744]: E1201 08:39:36.145970 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerName="dnsmasq-dns" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.146078 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerName="dnsmasq-dns" Dec 01 08:39:36 crc kubenswrapper[4744]: E1201 08:39:36.146160 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerName="init" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.146237 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerName="init" Dec 01 08:39:36 crc kubenswrapper[4744]: E1201 08:39:36.146338 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerName="dnsmasq-dns" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.146437 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerName="dnsmasq-dns" Dec 01 08:39:36 crc kubenswrapper[4744]: E1201 08:39:36.146528 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerName="init" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.146600 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerName="init" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.146907 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e6ef6aa-a18c-4dac-81c8-736c37f0e203" containerName="dnsmasq-dns" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.147025 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b4ff7ad-86db-42be-bb0f-9347c1c0494c" containerName="dnsmasq-dns" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.147868 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.151549 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.151919 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.152273 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.152280 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.184769 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz"] Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.231384 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.231918 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.232009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.232125 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz7gl\" (UniqueName: \"kubernetes.io/projected/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-kube-api-access-tz7gl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.334104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.334404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.334438 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.334468 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz7gl\" (UniqueName: \"kubernetes.io/projected/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-kube-api-access-tz7gl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.340137 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.341057 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.346070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.356270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz7gl\" (UniqueName: \"kubernetes.io/projected/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-kube-api-access-tz7gl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.485173 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.873061 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz"] Dec 01 08:39:36 crc kubenswrapper[4744]: W1201 08:39:36.878271 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75d5d4aa_8bf9_4042_b9e2_dfc02af5c1c6.slice/crio-95beb13a66fd404b53198b06a9be3d40ed5b0b16b627e223677dde72f145436b WatchSource:0}: Error finding container 95beb13a66fd404b53198b06a9be3d40ed5b0b16b627e223677dde72f145436b: Status 404 returned error can't find the container with id 95beb13a66fd404b53198b06a9be3d40ed5b0b16b627e223677dde72f145436b Dec 01 08:39:36 crc kubenswrapper[4744]: I1201 08:39:36.880747 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:39:37 crc kubenswrapper[4744]: I1201 08:39:37.263280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" event={"ID":"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6","Type":"ContainerStarted","Data":"95beb13a66fd404b53198b06a9be3d40ed5b0b16b627e223677dde72f145436b"} Dec 01 08:39:39 crc kubenswrapper[4744]: I1201 08:39:39.280606 4744 generic.go:334] "Generic (PLEG): container finished" podID="93e40ea3-265b-44a9-ada5-5b583201148d" containerID="ad795d047798f0653d4837b85e3555080dea221f31ae80a1b9683947819a30d7" exitCode=0 Dec 01 08:39:39 crc kubenswrapper[4744]: I1201 08:39:39.280702 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93e40ea3-265b-44a9-ada5-5b583201148d","Type":"ContainerDied","Data":"ad795d047798f0653d4837b85e3555080dea221f31ae80a1b9683947819a30d7"} Dec 01 08:39:40 crc kubenswrapper[4744]: I1201 08:39:40.297433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"93e40ea3-265b-44a9-ada5-5b583201148d","Type":"ContainerStarted","Data":"47381ad6254984dc8357d884e3b8ecb36267e8fdca0fbae03bce522ba824ffc4"} Dec 01 08:39:40 crc kubenswrapper[4744]: I1201 08:39:40.298295 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 08:39:40 crc kubenswrapper[4744]: I1201 08:39:40.322031 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.32200307 podStartE2EDuration="37.32200307s" podCreationTimestamp="2025-12-01 08:39:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:39:40.315866597 +0000 UTC m=+1312.304924528" watchObservedRunningTime="2025-12-01 08:39:40.32200307 +0000 UTC m=+1312.311060991" Dec 01 08:39:41 crc kubenswrapper[4744]: I1201 08:39:41.316152 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4a93bac-aabe-4109-950f-14cb6ac634dd" containerID="6230bf20d2289d169a27b1d6f81f3bf894ede30d1338c678de3ae2da07101643" exitCode=0 Dec 01 08:39:41 crc kubenswrapper[4744]: I1201 08:39:41.316256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4a93bac-aabe-4109-950f-14cb6ac634dd","Type":"ContainerDied","Data":"6230bf20d2289d169a27b1d6f81f3bf894ede30d1338c678de3ae2da07101643"} Dec 01 08:39:46 crc kubenswrapper[4744]: I1201 08:39:46.376572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e4a93bac-aabe-4109-950f-14cb6ac634dd","Type":"ContainerStarted","Data":"fd52052d665b7246fb3e61e11f2a212de3b39770196baae4ffd85e08168eeec6"} Dec 01 08:39:46 crc kubenswrapper[4744]: I1201 08:39:46.378779 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" event={"ID":"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6","Type":"ContainerStarted","Data":"197e7d088bd00d68183c1278fb33e33de6d7e9ddbb2ad22046bb49b686c0896f"} Dec 01 08:39:46 crc kubenswrapper[4744]: I1201 08:39:46.421682 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.421656275 podStartE2EDuration="41.421656275s" podCreationTimestamp="2025-12-01 08:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:39:46.409163122 +0000 UTC m=+1318.398221043" watchObservedRunningTime="2025-12-01 08:39:46.421656275 +0000 UTC m=+1318.410714236" Dec 01 08:39:46 crc kubenswrapper[4744]: I1201 08:39:46.435142 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" podStartSLOduration=1.363031175 podStartE2EDuration="10.435123534s" podCreationTimestamp="2025-12-01 08:39:36 +0000 UTC" firstStartedPulling="2025-12-01 08:39:36.880565391 +0000 UTC m=+1308.869623312" lastFinishedPulling="2025-12-01 08:39:45.95265775 +0000 UTC m=+1317.941715671" observedRunningTime="2025-12-01 08:39:46.428201549 +0000 UTC m=+1318.417259470" watchObservedRunningTime="2025-12-01 08:39:46.435123534 +0000 UTC m=+1318.424181445" Dec 01 08:39:53 crc kubenswrapper[4744]: I1201 08:39:53.550822 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 08:39:55 crc kubenswrapper[4744]: I1201 08:39:55.588101 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:39:57 crc kubenswrapper[4744]: I1201 08:39:57.502305 4744 generic.go:334] "Generic (PLEG): container finished" podID="75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" containerID="197e7d088bd00d68183c1278fb33e33de6d7e9ddbb2ad22046bb49b686c0896f" exitCode=0 Dec 01 08:39:57 crc kubenswrapper[4744]: I1201 08:39:57.502365 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" event={"ID":"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6","Type":"ContainerDied","Data":"197e7d088bd00d68183c1278fb33e33de6d7e9ddbb2ad22046bb49b686c0896f"} Dec 01 08:39:58 crc kubenswrapper[4744]: I1201 08:39:58.949238 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.088314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz7gl\" (UniqueName: \"kubernetes.io/projected/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-kube-api-access-tz7gl\") pod \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.088440 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-ssh-key\") pod \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.088544 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-repo-setup-combined-ca-bundle\") pod \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.088598 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-inventory\") pod \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\" (UID: \"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6\") " Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.093988 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" (UID: "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.099474 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-kube-api-access-tz7gl" (OuterVolumeSpecName: "kube-api-access-tz7gl") pod "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" (UID: "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6"). InnerVolumeSpecName "kube-api-access-tz7gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.118470 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-inventory" (OuterVolumeSpecName: "inventory") pod "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" (UID: "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.124741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" (UID: "75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.190634 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.190665 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.190678 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.190689 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz7gl\" (UniqueName: \"kubernetes.io/projected/75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6-kube-api-access-tz7gl\") on node \"crc\" DevicePath \"\"" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.524608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" event={"ID":"75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6","Type":"ContainerDied","Data":"95beb13a66fd404b53198b06a9be3d40ed5b0b16b627e223677dde72f145436b"} Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.524661 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95beb13a66fd404b53198b06a9be3d40ed5b0b16b627e223677dde72f145436b" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.524728 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.622061 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867"] Dec 01 08:39:59 crc kubenswrapper[4744]: E1201 08:39:59.622467 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.622482 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.622684 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.623312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.625477 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.626176 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.627548 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.628007 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.650792 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867"] Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.801260 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.801676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn6jn\" (UniqueName: \"kubernetes.io/projected/e59297bb-e715-4a12-b214-70d9aa5c4757-kube-api-access-jn6jn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.801969 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.904654 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.904913 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.905001 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn6jn\" (UniqueName: \"kubernetes.io/projected/e59297bb-e715-4a12-b214-70d9aa5c4757-kube-api-access-jn6jn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.908791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.909008 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.938253 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn6jn\" (UniqueName: \"kubernetes.io/projected/e59297bb-e715-4a12-b214-70d9aa5c4757-kube-api-access-jn6jn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mn867\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:39:59 crc kubenswrapper[4744]: I1201 08:39:59.946780 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:40:00 crc kubenswrapper[4744]: I1201 08:40:00.432379 4744 scope.go:117] "RemoveContainer" containerID="6d7a132512589b19cdbb339129b64fc6b547cdabf65f3f01841e3a7d19c8a2ea" Dec 01 08:40:00 crc kubenswrapper[4744]: I1201 08:40:00.531336 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867"] Dec 01 08:40:00 crc kubenswrapper[4744]: W1201 08:40:00.542384 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode59297bb_e715_4a12_b214_70d9aa5c4757.slice/crio-a8011bd6102f784472933c59e61f2525d5545370fa2015466e7a0c73cde041d9 WatchSource:0}: Error finding container a8011bd6102f784472933c59e61f2525d5545370fa2015466e7a0c73cde041d9: Status 404 returned error can't find the container with id a8011bd6102f784472933c59e61f2525d5545370fa2015466e7a0c73cde041d9 Dec 01 08:40:01 crc kubenswrapper[4744]: I1201 08:40:01.547444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" event={"ID":"e59297bb-e715-4a12-b214-70d9aa5c4757","Type":"ContainerStarted","Data":"52e2f5b07fa15a4007c1f7704d3ca95b700c66b346a8fc2935325d93ca3015ef"} Dec 01 08:40:01 crc kubenswrapper[4744]: I1201 08:40:01.547832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" event={"ID":"e59297bb-e715-4a12-b214-70d9aa5c4757","Type":"ContainerStarted","Data":"a8011bd6102f784472933c59e61f2525d5545370fa2015466e7a0c73cde041d9"} Dec 01 08:40:01 crc kubenswrapper[4744]: I1201 08:40:01.571862 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" podStartSLOduration=2.022215243 podStartE2EDuration="2.571833211s" podCreationTimestamp="2025-12-01 08:39:59 +0000 UTC" firstStartedPulling="2025-12-01 08:40:00.554490154 +0000 UTC m=+1332.543548075" lastFinishedPulling="2025-12-01 08:40:01.104108112 +0000 UTC m=+1333.093166043" observedRunningTime="2025-12-01 08:40:01.564870865 +0000 UTC m=+1333.553928786" watchObservedRunningTime="2025-12-01 08:40:01.571833211 +0000 UTC m=+1333.560891152" Dec 01 08:40:04 crc kubenswrapper[4744]: I1201 08:40:04.589183 4744 generic.go:334] "Generic (PLEG): container finished" podID="e59297bb-e715-4a12-b214-70d9aa5c4757" containerID="52e2f5b07fa15a4007c1f7704d3ca95b700c66b346a8fc2935325d93ca3015ef" exitCode=0 Dec 01 08:40:04 crc kubenswrapper[4744]: I1201 08:40:04.589392 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" event={"ID":"e59297bb-e715-4a12-b214-70d9aa5c4757","Type":"ContainerDied","Data":"52e2f5b07fa15a4007c1f7704d3ca95b700c66b346a8fc2935325d93ca3015ef"} Dec 01 08:40:05 crc kubenswrapper[4744]: I1201 08:40:05.590627 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.097987 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.243757 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-ssh-key\") pod \"e59297bb-e715-4a12-b214-70d9aa5c4757\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.244187 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn6jn\" (UniqueName: \"kubernetes.io/projected/e59297bb-e715-4a12-b214-70d9aa5c4757-kube-api-access-jn6jn\") pod \"e59297bb-e715-4a12-b214-70d9aa5c4757\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.244289 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-inventory\") pod \"e59297bb-e715-4a12-b214-70d9aa5c4757\" (UID: \"e59297bb-e715-4a12-b214-70d9aa5c4757\") " Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.249347 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e59297bb-e715-4a12-b214-70d9aa5c4757-kube-api-access-jn6jn" (OuterVolumeSpecName: "kube-api-access-jn6jn") pod "e59297bb-e715-4a12-b214-70d9aa5c4757" (UID: "e59297bb-e715-4a12-b214-70d9aa5c4757"). InnerVolumeSpecName "kube-api-access-jn6jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.279124 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-inventory" (OuterVolumeSpecName: "inventory") pod "e59297bb-e715-4a12-b214-70d9aa5c4757" (UID: "e59297bb-e715-4a12-b214-70d9aa5c4757"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.282529 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e59297bb-e715-4a12-b214-70d9aa5c4757" (UID: "e59297bb-e715-4a12-b214-70d9aa5c4757"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.346907 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.346938 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn6jn\" (UniqueName: \"kubernetes.io/projected/e59297bb-e715-4a12-b214-70d9aa5c4757-kube-api-access-jn6jn\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.346950 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e59297bb-e715-4a12-b214-70d9aa5c4757-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.617321 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" event={"ID":"e59297bb-e715-4a12-b214-70d9aa5c4757","Type":"ContainerDied","Data":"a8011bd6102f784472933c59e61f2525d5545370fa2015466e7a0c73cde041d9"} Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.617373 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8011bd6102f784472933c59e61f2525d5545370fa2015466e7a0c73cde041d9" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.617402 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mn867" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.713094 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c"] Dec 01 08:40:06 crc kubenswrapper[4744]: E1201 08:40:06.713505 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59297bb-e715-4a12-b214-70d9aa5c4757" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.713523 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59297bb-e715-4a12-b214-70d9aa5c4757" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.713762 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e59297bb-e715-4a12-b214-70d9aa5c4757" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.714624 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.716380 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.716506 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.716630 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.719730 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.723475 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c"] Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.857102 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.857151 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8bsq\" (UniqueName: \"kubernetes.io/projected/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-kube-api-access-w8bsq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.857509 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.857625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.959817 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.959863 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8bsq\" (UniqueName: \"kubernetes.io/projected/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-kube-api-access-w8bsq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.959954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.959986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.963845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.964267 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.964472 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:06 crc kubenswrapper[4744]: I1201 08:40:06.977138 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8bsq\" (UniqueName: \"kubernetes.io/projected/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-kube-api-access-w8bsq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:07 crc kubenswrapper[4744]: I1201 08:40:07.032852 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:40:07 crc kubenswrapper[4744]: I1201 08:40:07.556992 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c"] Dec 01 08:40:07 crc kubenswrapper[4744]: W1201 08:40:07.561757 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0546ddaa_13e4_481a_b1f1_a2fe6aa505f3.slice/crio-a23fb30ff477b357b7c0dcb0de9b939450418d410d38b4e8fd47a491de574a7f WatchSource:0}: Error finding container a23fb30ff477b357b7c0dcb0de9b939450418d410d38b4e8fd47a491de574a7f: Status 404 returned error can't find the container with id a23fb30ff477b357b7c0dcb0de9b939450418d410d38b4e8fd47a491de574a7f Dec 01 08:40:07 crc kubenswrapper[4744]: I1201 08:40:07.630358 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" event={"ID":"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3","Type":"ContainerStarted","Data":"a23fb30ff477b357b7c0dcb0de9b939450418d410d38b4e8fd47a491de574a7f"} Dec 01 08:40:08 crc kubenswrapper[4744]: I1201 08:40:08.642694 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" event={"ID":"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3","Type":"ContainerStarted","Data":"4b3e5e18431260c2553e7529afc8ccd4432bb975828ffd1b3df8457f12c0f68f"} Dec 01 08:40:08 crc kubenswrapper[4744]: I1201 08:40:08.661471 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" podStartSLOduration=1.963017877 podStartE2EDuration="2.661405619s" podCreationTimestamp="2025-12-01 08:40:06 +0000 UTC" firstStartedPulling="2025-12-01 08:40:07.565652261 +0000 UTC m=+1339.554710192" lastFinishedPulling="2025-12-01 08:40:08.264040023 +0000 UTC m=+1340.253097934" observedRunningTime="2025-12-01 08:40:08.657141288 +0000 UTC m=+1340.646199209" watchObservedRunningTime="2025-12-01 08:40:08.661405619 +0000 UTC m=+1340.650463550" Dec 01 08:40:18 crc kubenswrapper[4744]: I1201 08:40:18.634486 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:40:18 crc kubenswrapper[4744]: I1201 08:40:18.635042 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:40:48 crc kubenswrapper[4744]: I1201 08:40:48.634778 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:40:48 crc kubenswrapper[4744]: I1201 08:40:48.636345 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:41:00 crc kubenswrapper[4744]: I1201 08:41:00.566293 4744 scope.go:117] "RemoveContainer" containerID="b93baf2ded278cb1ba5a6756e714ffa5d0e48e1f5a0bbade056db46b165f01c0" Dec 01 08:41:00 crc kubenswrapper[4744]: I1201 08:41:00.598176 4744 scope.go:117] "RemoveContainer" containerID="f64967bbc5b21958514db37dd22f8dbe11703b724f23fb024f4d510562ac1cb0" Dec 01 08:41:18 crc kubenswrapper[4744]: I1201 08:41:18.634695 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:41:18 crc kubenswrapper[4744]: I1201 08:41:18.635210 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:41:18 crc kubenswrapper[4744]: I1201 08:41:18.635255 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:41:18 crc kubenswrapper[4744]: I1201 08:41:18.635850 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"610c82f121426bae487451dc7fd1be93ad3b6ec0bd33a7499656278834e982ab"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:41:18 crc kubenswrapper[4744]: I1201 08:41:18.635905 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://610c82f121426bae487451dc7fd1be93ad3b6ec0bd33a7499656278834e982ab" gracePeriod=600 Dec 01 08:41:19 crc kubenswrapper[4744]: I1201 08:41:19.348207 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="610c82f121426bae487451dc7fd1be93ad3b6ec0bd33a7499656278834e982ab" exitCode=0 Dec 01 08:41:19 crc kubenswrapper[4744]: I1201 08:41:19.348283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"610c82f121426bae487451dc7fd1be93ad3b6ec0bd33a7499656278834e982ab"} Dec 01 08:41:19 crc kubenswrapper[4744]: I1201 08:41:19.348576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd"} Dec 01 08:41:19 crc kubenswrapper[4744]: I1201 08:41:19.348598 4744 scope.go:117] "RemoveContainer" containerID="1e4f0e7898cde7a29868d7677f91a6fae395c27e58d4476c4df649466e26b91b" Dec 01 08:42:00 crc kubenswrapper[4744]: I1201 08:42:00.697300 4744 scope.go:117] "RemoveContainer" containerID="fad23623055fac729576a02472859863ce2d225d63941d7207a0100984ae31de" Dec 01 08:42:00 crc kubenswrapper[4744]: I1201 08:42:00.733165 4744 scope.go:117] "RemoveContainer" containerID="e3aaf2c7e38e8abea2025dc2c20bb09439a96aa7f14cd14295b1892504aaf946" Dec 01 08:42:00 crc kubenswrapper[4744]: I1201 08:42:00.761565 4744 scope.go:117] "RemoveContainer" containerID="b92d0cd7d77d5e49e122180332b578460f2a07dc39bfee62e14671c8a66b2d78" Dec 01 08:43:00 crc kubenswrapper[4744]: I1201 08:43:00.888570 4744 scope.go:117] "RemoveContainer" containerID="7e7449ab618120d846da616224c4a99796132a3829085f348a0e11ad06753fb2" Dec 01 08:43:00 crc kubenswrapper[4744]: I1201 08:43:00.916294 4744 scope.go:117] "RemoveContainer" containerID="4dfcdbfda44f684944eebb6bc027f16c1e974a30ab3daea1f18d5893d5a8dbca" Dec 01 08:43:00 crc kubenswrapper[4744]: I1201 08:43:00.940799 4744 scope.go:117] "RemoveContainer" containerID="279ddd346f8d6885a926bdd13cd2f10033ff65d0a75a9032142efb1674a25691" Dec 01 08:43:00 crc kubenswrapper[4744]: I1201 08:43:00.961285 4744 scope.go:117] "RemoveContainer" containerID="91d58797246780c729e3e1fdbfc9cc431f2165c5a32c748091f00c51e63af59c" Dec 01 08:43:19 crc kubenswrapper[4744]: I1201 08:43:19.545303 4744 generic.go:334] "Generic (PLEG): container finished" podID="0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" containerID="4b3e5e18431260c2553e7529afc8ccd4432bb975828ffd1b3df8457f12c0f68f" exitCode=0 Dec 01 08:43:19 crc kubenswrapper[4744]: I1201 08:43:19.545382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" event={"ID":"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3","Type":"ContainerDied","Data":"4b3e5e18431260c2553e7529afc8ccd4432bb975828ffd1b3df8457f12c0f68f"} Dec 01 08:43:20 crc kubenswrapper[4744]: I1201 08:43:20.994298 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.046752 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-inventory\") pod \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.046880 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-ssh-key\") pod \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.046927 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-bootstrap-combined-ca-bundle\") pod \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.047164 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8bsq\" (UniqueName: \"kubernetes.io/projected/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-kube-api-access-w8bsq\") pod \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\" (UID: \"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3\") " Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.052584 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" (UID: "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.054728 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-kube-api-access-w8bsq" (OuterVolumeSpecName: "kube-api-access-w8bsq") pod "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" (UID: "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3"). InnerVolumeSpecName "kube-api-access-w8bsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.080274 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" (UID: "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.092598 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-inventory" (OuterVolumeSpecName: "inventory") pod "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" (UID: "0546ddaa-13e4-481a-b1f1-a2fe6aa505f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.149441 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8bsq\" (UniqueName: \"kubernetes.io/projected/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-kube-api-access-w8bsq\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.149480 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.149490 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.149499 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0546ddaa-13e4-481a-b1f1-a2fe6aa505f3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.565271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" event={"ID":"0546ddaa-13e4-481a-b1f1-a2fe6aa505f3","Type":"ContainerDied","Data":"a23fb30ff477b357b7c0dcb0de9b939450418d410d38b4e8fd47a491de574a7f"} Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.565365 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a23fb30ff477b357b7c0dcb0de9b939450418d410d38b4e8fd47a491de574a7f" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.565451 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.661131 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl"] Dec 01 08:43:21 crc kubenswrapper[4744]: E1201 08:43:21.661607 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.661630 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.661883 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0546ddaa-13e4-481a-b1f1-a2fe6aa505f3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.662683 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.666778 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.667820 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.667862 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.669361 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.677366 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl"] Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.764835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.765059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.765221 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64x7c\" (UniqueName: \"kubernetes.io/projected/2b675722-4753-4de2-8bd8-aa37599eb53e-kube-api-access-64x7c\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.867983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.868119 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64x7c\" (UniqueName: \"kubernetes.io/projected/2b675722-4753-4de2-8bd8-aa37599eb53e-kube-api-access-64x7c\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.868189 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.875810 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.877469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.892259 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64x7c\" (UniqueName: \"kubernetes.io/projected/2b675722-4753-4de2-8bd8-aa37599eb53e-kube-api-access-64x7c\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:21 crc kubenswrapper[4744]: I1201 08:43:21.987107 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:43:22 crc kubenswrapper[4744]: I1201 08:43:22.498381 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl"] Dec 01 08:43:22 crc kubenswrapper[4744]: W1201 08:43:22.502430 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b675722_4753_4de2_8bd8_aa37599eb53e.slice/crio-a875ff2626c3a1026682a8bf93b6ca5014ede55fcda3bd2f12d4bd2ea966198a WatchSource:0}: Error finding container a875ff2626c3a1026682a8bf93b6ca5014ede55fcda3bd2f12d4bd2ea966198a: Status 404 returned error can't find the container with id a875ff2626c3a1026682a8bf93b6ca5014ede55fcda3bd2f12d4bd2ea966198a Dec 01 08:43:22 crc kubenswrapper[4744]: I1201 08:43:22.579571 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" event={"ID":"2b675722-4753-4de2-8bd8-aa37599eb53e","Type":"ContainerStarted","Data":"a875ff2626c3a1026682a8bf93b6ca5014ede55fcda3bd2f12d4bd2ea966198a"} Dec 01 08:43:23 crc kubenswrapper[4744]: I1201 08:43:23.590100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" event={"ID":"2b675722-4753-4de2-8bd8-aa37599eb53e","Type":"ContainerStarted","Data":"5ed59c29c086d6966c7739022fc532d9ca2761b5d1fdedb6386048cc17aa80eb"} Dec 01 08:43:23 crc kubenswrapper[4744]: I1201 08:43:23.613521 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" podStartSLOduration=1.7734294720000001 podStartE2EDuration="2.6135052s" podCreationTimestamp="2025-12-01 08:43:21 +0000 UTC" firstStartedPulling="2025-12-01 08:43:22.504708202 +0000 UTC m=+1534.493766123" lastFinishedPulling="2025-12-01 08:43:23.34478392 +0000 UTC m=+1535.333841851" observedRunningTime="2025-12-01 08:43:23.603467417 +0000 UTC m=+1535.592525348" watchObservedRunningTime="2025-12-01 08:43:23.6135052 +0000 UTC m=+1535.602563121" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.725612 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4wvnw"] Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.730344 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.741114 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wvnw"] Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.818308 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-utilities\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.818503 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r8jc\" (UniqueName: \"kubernetes.io/projected/5809a84a-20ac-4000-8891-bc0fc2d608e9-kube-api-access-4r8jc\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.818808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-catalog-content\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.920799 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-utilities\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.920884 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r8jc\" (UniqueName: \"kubernetes.io/projected/5809a84a-20ac-4000-8891-bc0fc2d608e9-kube-api-access-4r8jc\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.920955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-catalog-content\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.921435 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-catalog-content\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.921598 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-utilities\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:41 crc kubenswrapper[4744]: I1201 08:43:41.940727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r8jc\" (UniqueName: \"kubernetes.io/projected/5809a84a-20ac-4000-8891-bc0fc2d608e9-kube-api-access-4r8jc\") pod \"redhat-operators-4wvnw\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:42 crc kubenswrapper[4744]: I1201 08:43:42.055172 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:42 crc kubenswrapper[4744]: I1201 08:43:42.560466 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wvnw"] Dec 01 08:43:42 crc kubenswrapper[4744]: I1201 08:43:42.776648 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wvnw" event={"ID":"5809a84a-20ac-4000-8891-bc0fc2d608e9","Type":"ContainerStarted","Data":"011310ad243433a44567f36ade80ff675bf1822e87da64de09ab6ad881a5cf34"} Dec 01 08:43:43 crc kubenswrapper[4744]: I1201 08:43:43.790397 4744 generic.go:334] "Generic (PLEG): container finished" podID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerID="3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3" exitCode=0 Dec 01 08:43:43 crc kubenswrapper[4744]: I1201 08:43:43.790714 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wvnw" event={"ID":"5809a84a-20ac-4000-8891-bc0fc2d608e9","Type":"ContainerDied","Data":"3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3"} Dec 01 08:43:44 crc kubenswrapper[4744]: I1201 08:43:44.801751 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wvnw" event={"ID":"5809a84a-20ac-4000-8891-bc0fc2d608e9","Type":"ContainerStarted","Data":"032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094"} Dec 01 08:43:46 crc kubenswrapper[4744]: I1201 08:43:46.827138 4744 generic.go:334] "Generic (PLEG): container finished" podID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerID="032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094" exitCode=0 Dec 01 08:43:46 crc kubenswrapper[4744]: I1201 08:43:46.827211 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wvnw" event={"ID":"5809a84a-20ac-4000-8891-bc0fc2d608e9","Type":"ContainerDied","Data":"032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094"} Dec 01 08:43:48 crc kubenswrapper[4744]: I1201 08:43:48.635242 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:43:48 crc kubenswrapper[4744]: I1201 08:43:48.636842 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:43:48 crc kubenswrapper[4744]: I1201 08:43:48.852344 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wvnw" event={"ID":"5809a84a-20ac-4000-8891-bc0fc2d608e9","Type":"ContainerStarted","Data":"146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3"} Dec 01 08:43:48 crc kubenswrapper[4744]: I1201 08:43:48.877697 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4wvnw" podStartSLOduration=3.9325586059999997 podStartE2EDuration="7.877677442s" podCreationTimestamp="2025-12-01 08:43:41 +0000 UTC" firstStartedPulling="2025-12-01 08:43:43.793984479 +0000 UTC m=+1555.783042440" lastFinishedPulling="2025-12-01 08:43:47.739103355 +0000 UTC m=+1559.728161276" observedRunningTime="2025-12-01 08:43:48.870307027 +0000 UTC m=+1560.859364958" watchObservedRunningTime="2025-12-01 08:43:48.877677442 +0000 UTC m=+1560.866735363" Dec 01 08:43:52 crc kubenswrapper[4744]: I1201 08:43:52.055759 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:52 crc kubenswrapper[4744]: I1201 08:43:52.056387 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:43:53 crc kubenswrapper[4744]: I1201 08:43:53.113729 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4wvnw" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:53 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:53 crc kubenswrapper[4744]: > Dec 01 08:44:01 crc kubenswrapper[4744]: I1201 08:44:01.013354 4744 scope.go:117] "RemoveContainer" containerID="853334c399072bc499caa4ab92ef6484754bcfcdf9ad564e62a5b3663f302364" Dec 01 08:44:01 crc kubenswrapper[4744]: I1201 08:44:01.042643 4744 scope.go:117] "RemoveContainer" containerID="2557df3e2a35ca8c183f221af58573cea9a46f7dbcb7ebcf551e0a308534a886" Dec 01 08:44:02 crc kubenswrapper[4744]: I1201 08:44:02.124629 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:44:02 crc kubenswrapper[4744]: I1201 08:44:02.179906 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:44:02 crc kubenswrapper[4744]: I1201 08:44:02.369561 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wvnw"] Dec 01 08:44:03 crc kubenswrapper[4744]: I1201 08:44:03.994682 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4wvnw" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="registry-server" containerID="cri-o://146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3" gracePeriod=2 Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.512255 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.650231 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-catalog-content\") pod \"5809a84a-20ac-4000-8891-bc0fc2d608e9\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.650795 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-utilities\") pod \"5809a84a-20ac-4000-8891-bc0fc2d608e9\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.651071 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r8jc\" (UniqueName: \"kubernetes.io/projected/5809a84a-20ac-4000-8891-bc0fc2d608e9-kube-api-access-4r8jc\") pod \"5809a84a-20ac-4000-8891-bc0fc2d608e9\" (UID: \"5809a84a-20ac-4000-8891-bc0fc2d608e9\") " Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.651520 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-utilities" (OuterVolumeSpecName: "utilities") pod "5809a84a-20ac-4000-8891-bc0fc2d608e9" (UID: "5809a84a-20ac-4000-8891-bc0fc2d608e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.659074 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5809a84a-20ac-4000-8891-bc0fc2d608e9-kube-api-access-4r8jc" (OuterVolumeSpecName: "kube-api-access-4r8jc") pod "5809a84a-20ac-4000-8891-bc0fc2d608e9" (UID: "5809a84a-20ac-4000-8891-bc0fc2d608e9"). InnerVolumeSpecName "kube-api-access-4r8jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.753166 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r8jc\" (UniqueName: \"kubernetes.io/projected/5809a84a-20ac-4000-8891-bc0fc2d608e9-kube-api-access-4r8jc\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.753212 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.755066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5809a84a-20ac-4000-8891-bc0fc2d608e9" (UID: "5809a84a-20ac-4000-8891-bc0fc2d608e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:44:04 crc kubenswrapper[4744]: I1201 08:44:04.854630 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5809a84a-20ac-4000-8891-bc0fc2d608e9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.009507 4744 generic.go:334] "Generic (PLEG): container finished" podID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerID="146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3" exitCode=0 Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.009552 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wvnw" event={"ID":"5809a84a-20ac-4000-8891-bc0fc2d608e9","Type":"ContainerDied","Data":"146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3"} Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.009580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wvnw" event={"ID":"5809a84a-20ac-4000-8891-bc0fc2d608e9","Type":"ContainerDied","Data":"011310ad243433a44567f36ade80ff675bf1822e87da64de09ab6ad881a5cf34"} Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.009615 4744 scope.go:117] "RemoveContainer" containerID="146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.011178 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wvnw" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.038581 4744 scope.go:117] "RemoveContainer" containerID="032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.073198 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wvnw"] Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.092655 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4wvnw"] Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.097253 4744 scope.go:117] "RemoveContainer" containerID="3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.121915 4744 scope.go:117] "RemoveContainer" containerID="146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3" Dec 01 08:44:05 crc kubenswrapper[4744]: E1201 08:44:05.122331 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3\": container with ID starting with 146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3 not found: ID does not exist" containerID="146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.122367 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3"} err="failed to get container status \"146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3\": rpc error: code = NotFound desc = could not find container \"146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3\": container with ID starting with 146bd45fa2c13eb395747d8d1f54f808ddaed938d993e5a30dbb5c54986942c3 not found: ID does not exist" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.122390 4744 scope.go:117] "RemoveContainer" containerID="032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094" Dec 01 08:44:05 crc kubenswrapper[4744]: E1201 08:44:05.122812 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094\": container with ID starting with 032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094 not found: ID does not exist" containerID="032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.122858 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094"} err="failed to get container status \"032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094\": rpc error: code = NotFound desc = could not find container \"032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094\": container with ID starting with 032c4142995b7b56d1431e5a55e000aaaf729104ff07a45f52de77a9a319c094 not found: ID does not exist" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.122908 4744 scope.go:117] "RemoveContainer" containerID="3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3" Dec 01 08:44:05 crc kubenswrapper[4744]: E1201 08:44:05.123129 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3\": container with ID starting with 3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3 not found: ID does not exist" containerID="3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3" Dec 01 08:44:05 crc kubenswrapper[4744]: I1201 08:44:05.123177 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3"} err="failed to get container status \"3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3\": rpc error: code = NotFound desc = could not find container \"3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3\": container with ID starting with 3cf8535eb1285fde9674b8dbe366a6ddb9b888380684f6ef18b463dce939ced3 not found: ID does not exist" Dec 01 08:44:06 crc kubenswrapper[4744]: I1201 08:44:06.300244 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" path="/var/lib/kubelet/pods/5809a84a-20ac-4000-8891-bc0fc2d608e9/volumes" Dec 01 08:44:18 crc kubenswrapper[4744]: I1201 08:44:18.633829 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:44:18 crc kubenswrapper[4744]: I1201 08:44:18.635249 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.079621 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-b5w9k"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.094427 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4d38-account-create-update-d2vgj"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.105666 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8b45-account-create-update-tqk55"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.117308 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2c68m"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.128566 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-b5w9k"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.135892 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4d38-account-create-update-d2vgj"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.143074 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8b45-account-create-update-tqk55"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.150106 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2c68m"] Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.303941 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="164a9dc7-b73e-48e7-973a-a5adf61e37be" path="/var/lib/kubelet/pods/164a9dc7-b73e-48e7-973a-a5adf61e37be/volumes" Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.305768 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba566b4-fecf-4538-b02d-32b1173b0cd2" path="/var/lib/kubelet/pods/6ba566b4-fecf-4538-b02d-32b1173b0cd2/volumes" Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.307760 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3a278c5-97a6-4d6a-9a08-f770d6962f5d" path="/var/lib/kubelet/pods/c3a278c5-97a6-4d6a-9a08-f770d6962f5d/volumes" Dec 01 08:44:32 crc kubenswrapper[4744]: I1201 08:44:32.309720 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f627e1b2-a557-4531-bec4-a742a3f236d3" path="/var/lib/kubelet/pods/f627e1b2-a557-4531-bec4-a742a3f236d3/volumes" Dec 01 08:44:33 crc kubenswrapper[4744]: I1201 08:44:33.060779 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rz4d8"] Dec 01 08:44:33 crc kubenswrapper[4744]: I1201 08:44:33.078371 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-3282-account-create-update-q8hrf"] Dec 01 08:44:33 crc kubenswrapper[4744]: I1201 08:44:33.092482 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-3282-account-create-update-q8hrf"] Dec 01 08:44:33 crc kubenswrapper[4744]: I1201 08:44:33.104140 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rz4d8"] Dec 01 08:44:34 crc kubenswrapper[4744]: I1201 08:44:34.319458 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9476e9f-b34f-40b4-96d1-744e12762c72" path="/var/lib/kubelet/pods/a9476e9f-b34f-40b4-96d1-744e12762c72/volumes" Dec 01 08:44:34 crc kubenswrapper[4744]: I1201 08:44:34.321370 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b84980f5-3ba1-4b30-b572-4401e5487373" path="/var/lib/kubelet/pods/b84980f5-3ba1-4b30-b572-4401e5487373/volumes" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.709969 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2gx7j"] Dec 01 08:44:36 crc kubenswrapper[4744]: E1201 08:44:36.710938 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="extract-content" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.710961 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="extract-content" Dec 01 08:44:36 crc kubenswrapper[4744]: E1201 08:44:36.710977 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="extract-utilities" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.710988 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="extract-utilities" Dec 01 08:44:36 crc kubenswrapper[4744]: E1201 08:44:36.711013 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="registry-server" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.711025 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="registry-server" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.711371 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5809a84a-20ac-4000-8891-bc0fc2d608e9" containerName="registry-server" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.713663 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.722770 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gx7j"] Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.765624 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-catalog-content\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.765702 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-utilities\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.765808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkdgb\" (UniqueName: \"kubernetes.io/projected/0656478a-341e-431a-a3cc-5780025f8252-kube-api-access-qkdgb\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.866905 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-catalog-content\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.866951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-utilities\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.867014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkdgb\" (UniqueName: \"kubernetes.io/projected/0656478a-341e-431a-a3cc-5780025f8252-kube-api-access-qkdgb\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.867501 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-utilities\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.867597 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-catalog-content\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:36 crc kubenswrapper[4744]: I1201 08:44:36.891714 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkdgb\" (UniqueName: \"kubernetes.io/projected/0656478a-341e-431a-a3cc-5780025f8252-kube-api-access-qkdgb\") pod \"redhat-marketplace-2gx7j\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:37 crc kubenswrapper[4744]: I1201 08:44:37.065305 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:37 crc kubenswrapper[4744]: I1201 08:44:37.530363 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gx7j"] Dec 01 08:44:38 crc kubenswrapper[4744]: I1201 08:44:38.367835 4744 generic.go:334] "Generic (PLEG): container finished" podID="0656478a-341e-431a-a3cc-5780025f8252" containerID="96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad" exitCode=0 Dec 01 08:44:38 crc kubenswrapper[4744]: I1201 08:44:38.367948 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gx7j" event={"ID":"0656478a-341e-431a-a3cc-5780025f8252","Type":"ContainerDied","Data":"96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad"} Dec 01 08:44:38 crc kubenswrapper[4744]: I1201 08:44:38.368135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gx7j" event={"ID":"0656478a-341e-431a-a3cc-5780025f8252","Type":"ContainerStarted","Data":"5635fa6e8dfd688a4a06c59f3d6d50c122622890aad86e41271caa60ac827746"} Dec 01 08:44:38 crc kubenswrapper[4744]: I1201 08:44:38.370619 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:44:39 crc kubenswrapper[4744]: I1201 08:44:39.382639 4744 generic.go:334] "Generic (PLEG): container finished" podID="0656478a-341e-431a-a3cc-5780025f8252" containerID="ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2" exitCode=0 Dec 01 08:44:39 crc kubenswrapper[4744]: I1201 08:44:39.382707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gx7j" event={"ID":"0656478a-341e-431a-a3cc-5780025f8252","Type":"ContainerDied","Data":"ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2"} Dec 01 08:44:40 crc kubenswrapper[4744]: I1201 08:44:40.396743 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gx7j" event={"ID":"0656478a-341e-431a-a3cc-5780025f8252","Type":"ContainerStarted","Data":"5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c"} Dec 01 08:44:40 crc kubenswrapper[4744]: I1201 08:44:40.416740 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2gx7j" podStartSLOduration=2.808362086 podStartE2EDuration="4.416726459s" podCreationTimestamp="2025-12-01 08:44:36 +0000 UTC" firstStartedPulling="2025-12-01 08:44:38.370154541 +0000 UTC m=+1610.359212462" lastFinishedPulling="2025-12-01 08:44:39.978518904 +0000 UTC m=+1611.967576835" observedRunningTime="2025-12-01 08:44:40.412035592 +0000 UTC m=+1612.401093513" watchObservedRunningTime="2025-12-01 08:44:40.416726459 +0000 UTC m=+1612.405784380" Dec 01 08:44:47 crc kubenswrapper[4744]: I1201 08:44:47.066073 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:47 crc kubenswrapper[4744]: I1201 08:44:47.067059 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:47 crc kubenswrapper[4744]: I1201 08:44:47.142927 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:47 crc kubenswrapper[4744]: I1201 08:44:47.546459 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:47 crc kubenswrapper[4744]: I1201 08:44:47.601014 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gx7j"] Dec 01 08:44:48 crc kubenswrapper[4744]: I1201 08:44:48.634219 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:44:48 crc kubenswrapper[4744]: I1201 08:44:48.634274 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:44:48 crc kubenswrapper[4744]: I1201 08:44:48.634316 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:44:48 crc kubenswrapper[4744]: I1201 08:44:48.635175 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:44:48 crc kubenswrapper[4744]: I1201 08:44:48.635234 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" gracePeriod=600 Dec 01 08:44:48 crc kubenswrapper[4744]: E1201 08:44:48.759397 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:44:49 crc kubenswrapper[4744]: I1201 08:44:49.498439 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" exitCode=0 Dec 01 08:44:49 crc kubenswrapper[4744]: I1201 08:44:49.498441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd"} Dec 01 08:44:49 crc kubenswrapper[4744]: I1201 08:44:49.498817 4744 scope.go:117] "RemoveContainer" containerID="610c82f121426bae487451dc7fd1be93ad3b6ec0bd33a7499656278834e982ab" Dec 01 08:44:49 crc kubenswrapper[4744]: I1201 08:44:49.498937 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2gx7j" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="registry-server" containerID="cri-o://5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c" gracePeriod=2 Dec 01 08:44:49 crc kubenswrapper[4744]: I1201 08:44:49.499508 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:44:49 crc kubenswrapper[4744]: E1201 08:44:49.499763 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:44:49 crc kubenswrapper[4744]: I1201 08:44:49.959859 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.039603 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-catalog-content\") pod \"0656478a-341e-431a-a3cc-5780025f8252\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.039698 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkdgb\" (UniqueName: \"kubernetes.io/projected/0656478a-341e-431a-a3cc-5780025f8252-kube-api-access-qkdgb\") pod \"0656478a-341e-431a-a3cc-5780025f8252\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.039893 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-utilities\") pod \"0656478a-341e-431a-a3cc-5780025f8252\" (UID: \"0656478a-341e-431a-a3cc-5780025f8252\") " Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.040761 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-utilities" (OuterVolumeSpecName: "utilities") pod "0656478a-341e-431a-a3cc-5780025f8252" (UID: "0656478a-341e-431a-a3cc-5780025f8252"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.045849 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0656478a-341e-431a-a3cc-5780025f8252-kube-api-access-qkdgb" (OuterVolumeSpecName: "kube-api-access-qkdgb") pod "0656478a-341e-431a-a3cc-5780025f8252" (UID: "0656478a-341e-431a-a3cc-5780025f8252"). InnerVolumeSpecName "kube-api-access-qkdgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.060713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0656478a-341e-431a-a3cc-5780025f8252" (UID: "0656478a-341e-431a-a3cc-5780025f8252"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.142372 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.142446 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0656478a-341e-431a-a3cc-5780025f8252-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.142472 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkdgb\" (UniqueName: \"kubernetes.io/projected/0656478a-341e-431a-a3cc-5780025f8252-kube-api-access-qkdgb\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.513806 4744 generic.go:334] "Generic (PLEG): container finished" podID="0656478a-341e-431a-a3cc-5780025f8252" containerID="5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c" exitCode=0 Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.513875 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gx7j" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.513921 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gx7j" event={"ID":"0656478a-341e-431a-a3cc-5780025f8252","Type":"ContainerDied","Data":"5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c"} Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.514660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gx7j" event={"ID":"0656478a-341e-431a-a3cc-5780025f8252","Type":"ContainerDied","Data":"5635fa6e8dfd688a4a06c59f3d6d50c122622890aad86e41271caa60ac827746"} Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.514700 4744 scope.go:117] "RemoveContainer" containerID="5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.550387 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gx7j"] Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.554995 4744 scope.go:117] "RemoveContainer" containerID="ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.569617 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gx7j"] Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.586670 4744 scope.go:117] "RemoveContainer" containerID="96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.645687 4744 scope.go:117] "RemoveContainer" containerID="5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c" Dec 01 08:44:50 crc kubenswrapper[4744]: E1201 08:44:50.646391 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c\": container with ID starting with 5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c not found: ID does not exist" containerID="5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.646538 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c"} err="failed to get container status \"5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c\": rpc error: code = NotFound desc = could not find container \"5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c\": container with ID starting with 5b175c906755d71b514888b468afd5d8fc89619b3c6738a1801068912b5d870c not found: ID does not exist" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.646573 4744 scope.go:117] "RemoveContainer" containerID="ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2" Dec 01 08:44:50 crc kubenswrapper[4744]: E1201 08:44:50.647250 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2\": container with ID starting with ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2 not found: ID does not exist" containerID="ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.647329 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2"} err="failed to get container status \"ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2\": rpc error: code = NotFound desc = could not find container \"ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2\": container with ID starting with ee0ff864576cd0e294fe0fd64587462949e0c8e47033c80b91b7a56c0d366de2 not found: ID does not exist" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.647386 4744 scope.go:117] "RemoveContainer" containerID="96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad" Dec 01 08:44:50 crc kubenswrapper[4744]: E1201 08:44:50.649030 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad\": container with ID starting with 96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad not found: ID does not exist" containerID="96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad" Dec 01 08:44:50 crc kubenswrapper[4744]: I1201 08:44:50.649068 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad"} err="failed to get container status \"96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad\": rpc error: code = NotFound desc = could not find container \"96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad\": container with ID starting with 96c06411f6182de371aeb8571f63bf88af37be4b548589ecd62e6967b78a8dad not found: ID does not exist" Dec 01 08:44:52 crc kubenswrapper[4744]: I1201 08:44:52.313522 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0656478a-341e-431a-a3cc-5780025f8252" path="/var/lib/kubelet/pods/0656478a-341e-431a-a3cc-5780025f8252/volumes" Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.051845 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-b269-account-create-update-5rlsq"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.065834 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wg8gw"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.077559 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-b269-account-create-update-5rlsq"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.087637 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wg8gw"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.098279 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6364-account-create-update-9bhwp"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.107271 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6364-account-create-update-9bhwp"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.114039 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-gx62t"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.121112 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-4xvfq"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.127851 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-gx62t"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.134595 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-4xvfq"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.145647 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e015-account-create-update-m75kd"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.159623 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-fa0a-account-create-update-stqk2"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.170807 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-fa0a-account-create-update-stqk2"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.181541 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e015-account-create-update-m75kd"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.189135 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-9tggf"] Dec 01 08:44:57 crc kubenswrapper[4744]: I1201 08:44:57.196627 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-9tggf"] Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.300518 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="198a3f7b-ed68-4dee-92e1-9aa0e9362304" path="/var/lib/kubelet/pods/198a3f7b-ed68-4dee-92e1-9aa0e9362304/volumes" Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.302617 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4341bae7-8d88-4d04-8670-43125be4a7a3" path="/var/lib/kubelet/pods/4341bae7-8d88-4d04-8670-43125be4a7a3/volumes" Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.303879 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5076354c-7f79-490f-9ea7-c5d2dd60c545" path="/var/lib/kubelet/pods/5076354c-7f79-490f-9ea7-c5d2dd60c545/volumes" Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.306111 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fb2f8c0-0676-4f02-af2c-b4000c09ab26" path="/var/lib/kubelet/pods/5fb2f8c0-0676-4f02-af2c-b4000c09ab26/volumes" Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.308890 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e4d4551-0e28-432e-bd4e-e783133b0a4b" path="/var/lib/kubelet/pods/6e4d4551-0e28-432e-bd4e-e783133b0a4b/volumes" Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.310464 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ac6bc34-6d3b-4860-aab8-bdf00a485793" path="/var/lib/kubelet/pods/8ac6bc34-6d3b-4860-aab8-bdf00a485793/volumes" Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.311256 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b116a0ec-0413-4532-b820-181cbf9f3515" path="/var/lib/kubelet/pods/b116a0ec-0413-4532-b820-181cbf9f3515/volumes" Dec 01 08:44:58 crc kubenswrapper[4744]: I1201 08:44:58.312680 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed9aa660-b80d-4dd2-a150-fee9352d72b8" path="/var/lib/kubelet/pods/ed9aa660-b80d-4dd2-a150-fee9352d72b8/volumes" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.170206 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5"] Dec 01 08:45:00 crc kubenswrapper[4744]: E1201 08:45:00.171225 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="extract-utilities" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.171251 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="extract-utilities" Dec 01 08:45:00 crc kubenswrapper[4744]: E1201 08:45:00.171320 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="registry-server" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.171335 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="registry-server" Dec 01 08:45:00 crc kubenswrapper[4744]: E1201 08:45:00.171370 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="extract-content" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.171386 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="extract-content" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.171800 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0656478a-341e-431a-a3cc-5780025f8252" containerName="registry-server" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.172938 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.175935 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.176127 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.199550 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5"] Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.266717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fb169184-182d-416e-89e2-cafb9251f4d1-config-volume\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.266772 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbxms\" (UniqueName: \"kubernetes.io/projected/fb169184-182d-416e-89e2-cafb9251f4d1-kube-api-access-zbxms\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.266877 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fb169184-182d-416e-89e2-cafb9251f4d1-secret-volume\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.284999 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:45:00 crc kubenswrapper[4744]: E1201 08:45:00.285277 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.368318 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fb169184-182d-416e-89e2-cafb9251f4d1-config-volume\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.368515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbxms\" (UniqueName: \"kubernetes.io/projected/fb169184-182d-416e-89e2-cafb9251f4d1-kube-api-access-zbxms\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.368597 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fb169184-182d-416e-89e2-cafb9251f4d1-secret-volume\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.370668 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fb169184-182d-416e-89e2-cafb9251f4d1-config-volume\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.375617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fb169184-182d-416e-89e2-cafb9251f4d1-secret-volume\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.401653 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbxms\" (UniqueName: \"kubernetes.io/projected/fb169184-182d-416e-89e2-cafb9251f4d1-kube-api-access-zbxms\") pod \"collect-profiles-29409645-g7lr5\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:00 crc kubenswrapper[4744]: I1201 08:45:00.500631 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.064247 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5"] Dec 01 08:45:01 crc kubenswrapper[4744]: W1201 08:45:01.076324 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb169184_182d_416e_89e2_cafb9251f4d1.slice/crio-d360b51638faf9f5490f4a27331361160931d2f33f7d81c1b0a508caeba549b9 WatchSource:0}: Error finding container d360b51638faf9f5490f4a27331361160931d2f33f7d81c1b0a508caeba549b9: Status 404 returned error can't find the container with id d360b51638faf9f5490f4a27331361160931d2f33f7d81c1b0a508caeba549b9 Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.099563 4744 scope.go:117] "RemoveContainer" containerID="d3c1bff80bbaab51f7b3bb8f48ee7823284f295d0956c210eac4f8f02439dd15" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.137054 4744 scope.go:117] "RemoveContainer" containerID="7895b27b5816872126dcec77e8860c51572c626aad714162aca663072a7a3df7" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.177019 4744 scope.go:117] "RemoveContainer" containerID="244c4e59c0826358c4d1b6f4173ab80c21a61adfe61dcbdc9af0a3c6bc2c265f" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.209905 4744 scope.go:117] "RemoveContainer" containerID="3cb0c1e4ef01926601f9a5747ac84e703e156cf5c434af23492a3782035580d4" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.235762 4744 scope.go:117] "RemoveContainer" containerID="9ba871b26aa46f5224369b95b55187b2a66ebac58b9ec25ac192bd761432b827" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.268081 4744 scope.go:117] "RemoveContainer" containerID="46888f197156e6ef7c11d27d741085d1867e097fda7e53af0756b6301e4a36c4" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.296064 4744 scope.go:117] "RemoveContainer" containerID="a07782f241b0eb668936c74ff218d2d4b1d078ea4cf4a82909330cc1fcb52c2d" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.321340 4744 scope.go:117] "RemoveContainer" containerID="699bb5033f2dcaae20c9ce9cfa7d30a9bafa0539eb2d6ca6443120d168c80d6a" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.343346 4744 scope.go:117] "RemoveContainer" containerID="6f5660f26c5029a43db70500fcfb1dfeb7ad53b0b1d43841e145bf009afb4e06" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.365749 4744 scope.go:117] "RemoveContainer" containerID="982dfc2415048df980ec9f5d149660f2d3691dc57ad83c38d7fff673b2fea9b8" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.392202 4744 scope.go:117] "RemoveContainer" containerID="b883794a457fb19259ecefd6dac249e578dfe29a407fdb7f8f7ea0e04aba7f68" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.469884 4744 scope.go:117] "RemoveContainer" containerID="032b78dc265273587187f76090f5e446c42b62225f02ffb574c36c8b8411595a" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.538794 4744 scope.go:117] "RemoveContainer" containerID="44b0a76ff59530d05f1988f9403e2f890dd600e1d7a43a60e5cb7ffeff13911c" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.650965 4744 scope.go:117] "RemoveContainer" containerID="bc239bc0d3c935b996af080f41b06f11a86f65fabe629577114ed79e0af492b0" Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.660383 4744 generic.go:334] "Generic (PLEG): container finished" podID="fb169184-182d-416e-89e2-cafb9251f4d1" containerID="2d79de10d481a3d655b36f29f38936c7f27a4203fc8f6f987f1f81910e597c8e" exitCode=0 Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.660453 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" event={"ID":"fb169184-182d-416e-89e2-cafb9251f4d1","Type":"ContainerDied","Data":"2d79de10d481a3d655b36f29f38936c7f27a4203fc8f6f987f1f81910e597c8e"} Dec 01 08:45:01 crc kubenswrapper[4744]: I1201 08:45:01.660509 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" event={"ID":"fb169184-182d-416e-89e2-cafb9251f4d1","Type":"ContainerStarted","Data":"d360b51638faf9f5490f4a27331361160931d2f33f7d81c1b0a508caeba549b9"} Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.086935 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.130198 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbxms\" (UniqueName: \"kubernetes.io/projected/fb169184-182d-416e-89e2-cafb9251f4d1-kube-api-access-zbxms\") pod \"fb169184-182d-416e-89e2-cafb9251f4d1\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.130311 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fb169184-182d-416e-89e2-cafb9251f4d1-secret-volume\") pod \"fb169184-182d-416e-89e2-cafb9251f4d1\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.130478 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fb169184-182d-416e-89e2-cafb9251f4d1-config-volume\") pod \"fb169184-182d-416e-89e2-cafb9251f4d1\" (UID: \"fb169184-182d-416e-89e2-cafb9251f4d1\") " Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.131369 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb169184-182d-416e-89e2-cafb9251f4d1-config-volume" (OuterVolumeSpecName: "config-volume") pod "fb169184-182d-416e-89e2-cafb9251f4d1" (UID: "fb169184-182d-416e-89e2-cafb9251f4d1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.136468 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb169184-182d-416e-89e2-cafb9251f4d1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fb169184-182d-416e-89e2-cafb9251f4d1" (UID: "fb169184-182d-416e-89e2-cafb9251f4d1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.136673 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb169184-182d-416e-89e2-cafb9251f4d1-kube-api-access-zbxms" (OuterVolumeSpecName: "kube-api-access-zbxms") pod "fb169184-182d-416e-89e2-cafb9251f4d1" (UID: "fb169184-182d-416e-89e2-cafb9251f4d1"). InnerVolumeSpecName "kube-api-access-zbxms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.232745 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fb169184-182d-416e-89e2-cafb9251f4d1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.232775 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbxms\" (UniqueName: \"kubernetes.io/projected/fb169184-182d-416e-89e2-cafb9251f4d1-kube-api-access-zbxms\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.232789 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fb169184-182d-416e-89e2-cafb9251f4d1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.698023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" event={"ID":"fb169184-182d-416e-89e2-cafb9251f4d1","Type":"ContainerDied","Data":"d360b51638faf9f5490f4a27331361160931d2f33f7d81c1b0a508caeba549b9"} Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.698087 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d360b51638faf9f5490f4a27331361160931d2f33f7d81c1b0a508caeba549b9" Dec 01 08:45:03 crc kubenswrapper[4744]: I1201 08:45:03.698188 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5" Dec 01 08:45:04 crc kubenswrapper[4744]: I1201 08:45:04.056879 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jx6b8"] Dec 01 08:45:04 crc kubenswrapper[4744]: I1201 08:45:04.071069 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-d42bc"] Dec 01 08:45:04 crc kubenswrapper[4744]: I1201 08:45:04.081058 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-d42bc"] Dec 01 08:45:04 crc kubenswrapper[4744]: I1201 08:45:04.089774 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jx6b8"] Dec 01 08:45:04 crc kubenswrapper[4744]: I1201 08:45:04.306238 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08cb1243-7645-41a5-b5d2-ea22230908c7" path="/var/lib/kubelet/pods/08cb1243-7645-41a5-b5d2-ea22230908c7/volumes" Dec 01 08:45:04 crc kubenswrapper[4744]: I1201 08:45:04.309994 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d725b87c-b27b-414d-824d-e9f9f3172d69" path="/var/lib/kubelet/pods/d725b87c-b27b-414d-824d-e9f9f3172d69/volumes" Dec 01 08:45:11 crc kubenswrapper[4744]: I1201 08:45:11.286984 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:45:11 crc kubenswrapper[4744]: E1201 08:45:11.287996 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:45:15 crc kubenswrapper[4744]: I1201 08:45:15.834177 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b675722-4753-4de2-8bd8-aa37599eb53e" containerID="5ed59c29c086d6966c7739022fc532d9ca2761b5d1fdedb6386048cc17aa80eb" exitCode=0 Dec 01 08:45:15 crc kubenswrapper[4744]: I1201 08:45:15.834235 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" event={"ID":"2b675722-4753-4de2-8bd8-aa37599eb53e","Type":"ContainerDied","Data":"5ed59c29c086d6966c7739022fc532d9ca2761b5d1fdedb6386048cc17aa80eb"} Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.319500 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.431789 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64x7c\" (UniqueName: \"kubernetes.io/projected/2b675722-4753-4de2-8bd8-aa37599eb53e-kube-api-access-64x7c\") pod \"2b675722-4753-4de2-8bd8-aa37599eb53e\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.431950 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-inventory\") pod \"2b675722-4753-4de2-8bd8-aa37599eb53e\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.431980 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-ssh-key\") pod \"2b675722-4753-4de2-8bd8-aa37599eb53e\" (UID: \"2b675722-4753-4de2-8bd8-aa37599eb53e\") " Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.439379 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b675722-4753-4de2-8bd8-aa37599eb53e-kube-api-access-64x7c" (OuterVolumeSpecName: "kube-api-access-64x7c") pod "2b675722-4753-4de2-8bd8-aa37599eb53e" (UID: "2b675722-4753-4de2-8bd8-aa37599eb53e"). InnerVolumeSpecName "kube-api-access-64x7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.476061 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2b675722-4753-4de2-8bd8-aa37599eb53e" (UID: "2b675722-4753-4de2-8bd8-aa37599eb53e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.488723 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-inventory" (OuterVolumeSpecName: "inventory") pod "2b675722-4753-4de2-8bd8-aa37599eb53e" (UID: "2b675722-4753-4de2-8bd8-aa37599eb53e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.534748 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.534809 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b675722-4753-4de2-8bd8-aa37599eb53e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.534822 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64x7c\" (UniqueName: \"kubernetes.io/projected/2b675722-4753-4de2-8bd8-aa37599eb53e-kube-api-access-64x7c\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.863729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" event={"ID":"2b675722-4753-4de2-8bd8-aa37599eb53e","Type":"ContainerDied","Data":"a875ff2626c3a1026682a8bf93b6ca5014ede55fcda3bd2f12d4bd2ea966198a"} Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.863788 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a875ff2626c3a1026682a8bf93b6ca5014ede55fcda3bd2f12d4bd2ea966198a" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.863829 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.985918 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766"] Dec 01 08:45:17 crc kubenswrapper[4744]: E1201 08:45:17.986288 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b675722-4753-4de2-8bd8-aa37599eb53e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.986305 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b675722-4753-4de2-8bd8-aa37599eb53e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 08:45:17 crc kubenswrapper[4744]: E1201 08:45:17.986342 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb169184-182d-416e-89e2-cafb9251f4d1" containerName="collect-profiles" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.986350 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb169184-182d-416e-89e2-cafb9251f4d1" containerName="collect-profiles" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.986532 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b675722-4753-4de2-8bd8-aa37599eb53e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.986544 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb169184-182d-416e-89e2-cafb9251f4d1" containerName="collect-profiles" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.987241 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.990499 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.991058 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.991509 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:45:17 crc kubenswrapper[4744]: I1201 08:45:17.991882 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.004259 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766"] Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.147486 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.147595 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.147615 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92t4k\" (UniqueName: \"kubernetes.io/projected/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-kube-api-access-92t4k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.249485 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.249578 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.249600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92t4k\" (UniqueName: \"kubernetes.io/projected/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-kube-api-access-92t4k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.256654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.260795 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.277708 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92t4k\" (UniqueName: \"kubernetes.io/projected/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-kube-api-access-92t4k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4766\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.314959 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:45:18 crc kubenswrapper[4744]: I1201 08:45:18.898546 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766"] Dec 01 08:45:19 crc kubenswrapper[4744]: I1201 08:45:19.905012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" event={"ID":"1f3aaa29-4bf7-4f79-aae9-7e90d131703d","Type":"ContainerStarted","Data":"74d99e944d94584aac64ba559baabe965764ecb30f7511decc1afe907cca4f1d"} Dec 01 08:45:20 crc kubenswrapper[4744]: I1201 08:45:20.921841 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" event={"ID":"1f3aaa29-4bf7-4f79-aae9-7e90d131703d","Type":"ContainerStarted","Data":"b64ffae0eefde203842acd7ba943de4aeb49cbc2bf883b803034e409ab4c823d"} Dec 01 08:45:20 crc kubenswrapper[4744]: I1201 08:45:20.949245 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" podStartSLOduration=3.169733262 podStartE2EDuration="3.949219983s" podCreationTimestamp="2025-12-01 08:45:17 +0000 UTC" firstStartedPulling="2025-12-01 08:45:18.893913332 +0000 UTC m=+1650.882971253" lastFinishedPulling="2025-12-01 08:45:19.673400053 +0000 UTC m=+1651.662457974" observedRunningTime="2025-12-01 08:45:20.945716641 +0000 UTC m=+1652.934774602" watchObservedRunningTime="2025-12-01 08:45:20.949219983 +0000 UTC m=+1652.938277934" Dec 01 08:45:25 crc kubenswrapper[4744]: I1201 08:45:25.285005 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:45:25 crc kubenswrapper[4744]: E1201 08:45:25.286855 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:45:36 crc kubenswrapper[4744]: I1201 08:45:36.285754 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:45:36 crc kubenswrapper[4744]: E1201 08:45:36.286484 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:45:38 crc kubenswrapper[4744]: I1201 08:45:38.039157 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-5k8xk"] Dec 01 08:45:38 crc kubenswrapper[4744]: I1201 08:45:38.048009 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-5k8xk"] Dec 01 08:45:38 crc kubenswrapper[4744]: I1201 08:45:38.303037 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d51f040d-3a61-47f1-8699-c88fbaabdee5" path="/var/lib/kubelet/pods/d51f040d-3a61-47f1-8699-c88fbaabdee5/volumes" Dec 01 08:45:43 crc kubenswrapper[4744]: I1201 08:45:43.049717 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-bzxk9"] Dec 01 08:45:43 crc kubenswrapper[4744]: I1201 08:45:43.064695 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xdnlb"] Dec 01 08:45:43 crc kubenswrapper[4744]: I1201 08:45:43.076832 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xdnlb"] Dec 01 08:45:43 crc kubenswrapper[4744]: I1201 08:45:43.087014 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-bzxk9"] Dec 01 08:45:44 crc kubenswrapper[4744]: I1201 08:45:44.299172 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95085c33-3a8b-4af5-aaef-bd718dabcf6f" path="/var/lib/kubelet/pods/95085c33-3a8b-4af5-aaef-bd718dabcf6f/volumes" Dec 01 08:45:44 crc kubenswrapper[4744]: I1201 08:45:44.301191 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf20b7a9-6976-4f37-8cfc-f3c67d2057c2" path="/var/lib/kubelet/pods/cf20b7a9-6976-4f37-8cfc-f3c67d2057c2/volumes" Dec 01 08:45:47 crc kubenswrapper[4744]: I1201 08:45:47.285930 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:45:47 crc kubenswrapper[4744]: E1201 08:45:47.286477 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:45:58 crc kubenswrapper[4744]: I1201 08:45:58.033345 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-8mwhk"] Dec 01 08:45:58 crc kubenswrapper[4744]: I1201 08:45:58.042659 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-8mwhk"] Dec 01 08:45:58 crc kubenswrapper[4744]: I1201 08:45:58.291976 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:45:58 crc kubenswrapper[4744]: E1201 08:45:58.293266 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:45:58 crc kubenswrapper[4744]: I1201 08:45:58.301585 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed2a24ae-749f-4faf-b077-060df14120bb" path="/var/lib/kubelet/pods/ed2a24ae-749f-4faf-b077-060df14120bb/volumes" Dec 01 08:46:00 crc kubenswrapper[4744]: I1201 08:46:00.050350 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-x664l"] Dec 01 08:46:00 crc kubenswrapper[4744]: I1201 08:46:00.062883 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-t2mxc"] Dec 01 08:46:00 crc kubenswrapper[4744]: I1201 08:46:00.074065 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-x664l"] Dec 01 08:46:00 crc kubenswrapper[4744]: I1201 08:46:00.094065 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-t2mxc"] Dec 01 08:46:00 crc kubenswrapper[4744]: I1201 08:46:00.299714 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab3b54ed-348a-4314-8e92-3e977ee8f1ff" path="/var/lib/kubelet/pods/ab3b54ed-348a-4314-8e92-3e977ee8f1ff/volumes" Dec 01 08:46:00 crc kubenswrapper[4744]: I1201 08:46:00.301839 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3718521-a22d-4350-aa32-98987bfe5632" path="/var/lib/kubelet/pods/d3718521-a22d-4350-aa32-98987bfe5632/volumes" Dec 01 08:46:01 crc kubenswrapper[4744]: I1201 08:46:01.996057 4744 scope.go:117] "RemoveContainer" containerID="803acab525a520838fffbec016e53bdc2404fe4cd412bcc287c30998b04dc8dd" Dec 01 08:46:02 crc kubenswrapper[4744]: I1201 08:46:02.052330 4744 scope.go:117] "RemoveContainer" containerID="03c20540d1d86beb58aaacb11728a294a18816f02865bcb699d8d6184cf66c6b" Dec 01 08:46:02 crc kubenswrapper[4744]: I1201 08:46:02.135684 4744 scope.go:117] "RemoveContainer" containerID="6ab49ff608f7526926bb7b38c3b848d5dea0d006ccc3a811638cf2b9c3a16ca1" Dec 01 08:46:02 crc kubenswrapper[4744]: I1201 08:46:02.196473 4744 scope.go:117] "RemoveContainer" containerID="772e5a2097a6c125df66c7652c0e9b97632c6ed6ab9088063178831126e57bab" Dec 01 08:46:02 crc kubenswrapper[4744]: I1201 08:46:02.243933 4744 scope.go:117] "RemoveContainer" containerID="cf486ae0b0a4c2c372ea294f78931e59ebf5398ab8cd5b7a34c4a3a403a4032a" Dec 01 08:46:02 crc kubenswrapper[4744]: I1201 08:46:02.308108 4744 scope.go:117] "RemoveContainer" containerID="3c4190b993a15f411262d140d6dc55defae79c915dc8bbfafcb13a406c81c171" Dec 01 08:46:02 crc kubenswrapper[4744]: I1201 08:46:02.355917 4744 scope.go:117] "RemoveContainer" containerID="a8b3f04302e8835bc76b504b46fc1c5e954d957f2e75d1b3ece12306aae68748" Dec 01 08:46:02 crc kubenswrapper[4744]: I1201 08:46:02.402995 4744 scope.go:117] "RemoveContainer" containerID="f601837c12e23fdf5cb5eeecf24667a4e5c865a0178787fd7f6e26a81fa80f09" Dec 01 08:46:09 crc kubenswrapper[4744]: I1201 08:46:09.284960 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:46:09 crc kubenswrapper[4744]: E1201 08:46:09.285727 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:46:21 crc kubenswrapper[4744]: I1201 08:46:21.286216 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:46:21 crc kubenswrapper[4744]: E1201 08:46:21.288923 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:46:33 crc kubenswrapper[4744]: I1201 08:46:33.284920 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:46:33 crc kubenswrapper[4744]: E1201 08:46:33.285761 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:46:37 crc kubenswrapper[4744]: I1201 08:46:37.752980 4744 generic.go:334] "Generic (PLEG): container finished" podID="1f3aaa29-4bf7-4f79-aae9-7e90d131703d" containerID="b64ffae0eefde203842acd7ba943de4aeb49cbc2bf883b803034e409ab4c823d" exitCode=0 Dec 01 08:46:37 crc kubenswrapper[4744]: I1201 08:46:37.753065 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" event={"ID":"1f3aaa29-4bf7-4f79-aae9-7e90d131703d","Type":"ContainerDied","Data":"b64ffae0eefde203842acd7ba943de4aeb49cbc2bf883b803034e409ab4c823d"} Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.242081 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.408131 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-inventory\") pod \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.408515 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-ssh-key\") pod \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.408603 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92t4k\" (UniqueName: \"kubernetes.io/projected/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-kube-api-access-92t4k\") pod \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\" (UID: \"1f3aaa29-4bf7-4f79-aae9-7e90d131703d\") " Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.422572 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-kube-api-access-92t4k" (OuterVolumeSpecName: "kube-api-access-92t4k") pod "1f3aaa29-4bf7-4f79-aae9-7e90d131703d" (UID: "1f3aaa29-4bf7-4f79-aae9-7e90d131703d"). InnerVolumeSpecName "kube-api-access-92t4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.443110 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1f3aaa29-4bf7-4f79-aae9-7e90d131703d" (UID: "1f3aaa29-4bf7-4f79-aae9-7e90d131703d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.463229 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-inventory" (OuterVolumeSpecName: "inventory") pod "1f3aaa29-4bf7-4f79-aae9-7e90d131703d" (UID: "1f3aaa29-4bf7-4f79-aae9-7e90d131703d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.511222 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.511267 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.511280 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92t4k\" (UniqueName: \"kubernetes.io/projected/1f3aaa29-4bf7-4f79-aae9-7e90d131703d-kube-api-access-92t4k\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.776830 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" event={"ID":"1f3aaa29-4bf7-4f79-aae9-7e90d131703d","Type":"ContainerDied","Data":"74d99e944d94584aac64ba559baabe965764ecb30f7511decc1afe907cca4f1d"} Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.776883 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74d99e944d94584aac64ba559baabe965764ecb30f7511decc1afe907cca4f1d" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.778132 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4766" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.937966 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt"] Dec 01 08:46:39 crc kubenswrapper[4744]: E1201 08:46:39.938448 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f3aaa29-4bf7-4f79-aae9-7e90d131703d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.938468 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f3aaa29-4bf7-4f79-aae9-7e90d131703d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.938731 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f3aaa29-4bf7-4f79-aae9-7e90d131703d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.939533 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.944630 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.945318 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.947399 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt"] Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.948662 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:46:39 crc kubenswrapper[4744]: I1201 08:46:39.948754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.122287 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.122350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwddn\" (UniqueName: \"kubernetes.io/projected/25cf1554-6aff-4169-ae0e-a6b6979ad45e-kube-api-access-zwddn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.122870 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.224599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.224750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.224785 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwddn\" (UniqueName: \"kubernetes.io/projected/25cf1554-6aff-4169-ae0e-a6b6979ad45e-kube-api-access-zwddn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.228666 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.231923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.253769 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwddn\" (UniqueName: \"kubernetes.io/projected/25cf1554-6aff-4169-ae0e-a6b6979ad45e-kube-api-access-zwddn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.255759 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:40 crc kubenswrapper[4744]: I1201 08:46:40.856006 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt"] Dec 01 08:46:41 crc kubenswrapper[4744]: I1201 08:46:41.796714 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" event={"ID":"25cf1554-6aff-4169-ae0e-a6b6979ad45e","Type":"ContainerStarted","Data":"4e257af249eaceef008775d290ae17fe09e170d23fc5696d89a9671107011533"} Dec 01 08:46:41 crc kubenswrapper[4744]: I1201 08:46:41.797304 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" event={"ID":"25cf1554-6aff-4169-ae0e-a6b6979ad45e","Type":"ContainerStarted","Data":"c806908946809bc2004d61f83a8b48dd3c6275763e5f78d1a0a90f4e47c148c5"} Dec 01 08:46:41 crc kubenswrapper[4744]: I1201 08:46:41.820659 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" podStartSLOduration=2.304796704 podStartE2EDuration="2.82063837s" podCreationTimestamp="2025-12-01 08:46:39 +0000 UTC" firstStartedPulling="2025-12-01 08:46:40.863322331 +0000 UTC m=+1732.852380252" lastFinishedPulling="2025-12-01 08:46:41.379163997 +0000 UTC m=+1733.368221918" observedRunningTime="2025-12-01 08:46:41.811640173 +0000 UTC m=+1733.800698104" watchObservedRunningTime="2025-12-01 08:46:41.82063837 +0000 UTC m=+1733.809696301" Dec 01 08:46:43 crc kubenswrapper[4744]: I1201 08:46:43.064701 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-k46jn"] Dec 01 08:46:43 crc kubenswrapper[4744]: I1201 08:46:43.073478 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-k46jn"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.038042 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-sslnq"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.051309 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f431-account-create-update-lvs6w"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.065499 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8cec-account-create-update-xlrzh"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.077827 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f431-account-create-update-lvs6w"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.089474 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-sslnq"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.096749 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8cec-account-create-update-xlrzh"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.106659 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-fe16-account-create-update-h8bdl"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.116875 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-fe16-account-create-update-h8bdl"] Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.286039 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:46:44 crc kubenswrapper[4744]: E1201 08:46:44.286389 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.297370 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59632a6c-e03f-45dd-a08a-561603cb97b5" path="/var/lib/kubelet/pods/59632a6c-e03f-45dd-a08a-561603cb97b5/volumes" Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.299136 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65092f5c-34d0-4eaa-9f53-dea23cc222d4" path="/var/lib/kubelet/pods/65092f5c-34d0-4eaa-9f53-dea23cc222d4/volumes" Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.300511 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7230ec92-1f41-491b-a2b5-23aa29cf4d27" path="/var/lib/kubelet/pods/7230ec92-1f41-491b-a2b5-23aa29cf4d27/volumes" Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.301930 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80d42461-4c7a-4e38-8ac7-1261d5fcb37a" path="/var/lib/kubelet/pods/80d42461-4c7a-4e38-8ac7-1261d5fcb37a/volumes" Dec 01 08:46:44 crc kubenswrapper[4744]: I1201 08:46:44.304034 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5172715-95f9-48a5-bf83-8fbe12093298" path="/var/lib/kubelet/pods/e5172715-95f9-48a5-bf83-8fbe12093298/volumes" Dec 01 08:46:45 crc kubenswrapper[4744]: I1201 08:46:45.041911 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5fv45"] Dec 01 08:46:45 crc kubenswrapper[4744]: I1201 08:46:45.059594 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5fv45"] Dec 01 08:46:46 crc kubenswrapper[4744]: I1201 08:46:46.302595 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="606662e8-bd00-4a68-ba9f-26b9d93d175b" path="/var/lib/kubelet/pods/606662e8-bd00-4a68-ba9f-26b9d93d175b/volumes" Dec 01 08:46:47 crc kubenswrapper[4744]: I1201 08:46:47.855845 4744 generic.go:334] "Generic (PLEG): container finished" podID="25cf1554-6aff-4169-ae0e-a6b6979ad45e" containerID="4e257af249eaceef008775d290ae17fe09e170d23fc5696d89a9671107011533" exitCode=0 Dec 01 08:46:47 crc kubenswrapper[4744]: I1201 08:46:47.855896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" event={"ID":"25cf1554-6aff-4169-ae0e-a6b6979ad45e","Type":"ContainerDied","Data":"4e257af249eaceef008775d290ae17fe09e170d23fc5696d89a9671107011533"} Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.372731 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.521114 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-inventory\") pod \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.521988 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-ssh-key\") pod \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.522086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwddn\" (UniqueName: \"kubernetes.io/projected/25cf1554-6aff-4169-ae0e-a6b6979ad45e-kube-api-access-zwddn\") pod \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\" (UID: \"25cf1554-6aff-4169-ae0e-a6b6979ad45e\") " Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.541754 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25cf1554-6aff-4169-ae0e-a6b6979ad45e-kube-api-access-zwddn" (OuterVolumeSpecName: "kube-api-access-zwddn") pod "25cf1554-6aff-4169-ae0e-a6b6979ad45e" (UID: "25cf1554-6aff-4169-ae0e-a6b6979ad45e"). InnerVolumeSpecName "kube-api-access-zwddn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.564278 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-inventory" (OuterVolumeSpecName: "inventory") pod "25cf1554-6aff-4169-ae0e-a6b6979ad45e" (UID: "25cf1554-6aff-4169-ae0e-a6b6979ad45e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.577788 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "25cf1554-6aff-4169-ae0e-a6b6979ad45e" (UID: "25cf1554-6aff-4169-ae0e-a6b6979ad45e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.624256 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.624283 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwddn\" (UniqueName: \"kubernetes.io/projected/25cf1554-6aff-4169-ae0e-a6b6979ad45e-kube-api-access-zwddn\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.624291 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25cf1554-6aff-4169-ae0e-a6b6979ad45e-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.876968 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" event={"ID":"25cf1554-6aff-4169-ae0e-a6b6979ad45e","Type":"ContainerDied","Data":"c806908946809bc2004d61f83a8b48dd3c6275763e5f78d1a0a90f4e47c148c5"} Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.877024 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.877026 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c806908946809bc2004d61f83a8b48dd3c6275763e5f78d1a0a90f4e47c148c5" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.957771 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk"] Dec 01 08:46:49 crc kubenswrapper[4744]: E1201 08:46:49.958152 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cf1554-6aff-4169-ae0e-a6b6979ad45e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.958170 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cf1554-6aff-4169-ae0e-a6b6979ad45e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.958444 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="25cf1554-6aff-4169-ae0e-a6b6979ad45e" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.959289 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.961710 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.961989 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.962277 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.962568 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:46:49 crc kubenswrapper[4744]: I1201 08:46:49.973523 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk"] Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.133542 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wnsk\" (UniqueName: \"kubernetes.io/projected/b2c369ea-0c26-424a-af22-996d64ecf25d-kube-api-access-9wnsk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.133799 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.133841 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.235107 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.235153 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.235284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wnsk\" (UniqueName: \"kubernetes.io/projected/b2c369ea-0c26-424a-af22-996d64ecf25d-kube-api-access-9wnsk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.238791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.246047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.254530 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wnsk\" (UniqueName: \"kubernetes.io/projected/b2c369ea-0c26-424a-af22-996d64ecf25d-kube-api-access-9wnsk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fttwk\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.291766 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.846993 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk"] Dec 01 08:46:50 crc kubenswrapper[4744]: I1201 08:46:50.904593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" event={"ID":"b2c369ea-0c26-424a-af22-996d64ecf25d","Type":"ContainerStarted","Data":"545b89f8c73334fa8ad097918981dc0ba085f31b376889abe09deae528588b7b"} Dec 01 08:46:51 crc kubenswrapper[4744]: I1201 08:46:51.912464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" event={"ID":"b2c369ea-0c26-424a-af22-996d64ecf25d","Type":"ContainerStarted","Data":"ca4e14d9f6a46a57a7f41ec8c6e5ee035e997153a875e0a32a5c2c5c3ae4da6e"} Dec 01 08:46:51 crc kubenswrapper[4744]: I1201 08:46:51.932069 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" podStartSLOduration=2.415180328 podStartE2EDuration="2.932044873s" podCreationTimestamp="2025-12-01 08:46:49 +0000 UTC" firstStartedPulling="2025-12-01 08:46:50.85686414 +0000 UTC m=+1742.845922061" lastFinishedPulling="2025-12-01 08:46:51.373728685 +0000 UTC m=+1743.362786606" observedRunningTime="2025-12-01 08:46:51.92774466 +0000 UTC m=+1743.916802621" watchObservedRunningTime="2025-12-01 08:46:51.932044873 +0000 UTC m=+1743.921102794" Dec 01 08:46:57 crc kubenswrapper[4744]: I1201 08:46:57.284967 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:46:57 crc kubenswrapper[4744]: E1201 08:46:57.285805 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:47:02 crc kubenswrapper[4744]: I1201 08:47:02.650042 4744 scope.go:117] "RemoveContainer" containerID="d66a177ed2e123f2913d76f6d902840e58acedd6aa55d4d73aa798862db4035d" Dec 01 08:47:02 crc kubenswrapper[4744]: I1201 08:47:02.682984 4744 scope.go:117] "RemoveContainer" containerID="78e0eae410b90b6f7bcfb7f22d1b6e9e69f6b122e21b37edf7d6df78fe8e20d1" Dec 01 08:47:02 crc kubenswrapper[4744]: I1201 08:47:02.752926 4744 scope.go:117] "RemoveContainer" containerID="7f2960ab28ffbc55a4e737a7b3e8ba326b0ddfa06f7cac5faa617e8eb314a7e8" Dec 01 08:47:02 crc kubenswrapper[4744]: I1201 08:47:02.788203 4744 scope.go:117] "RemoveContainer" containerID="d8419e9a848a43bb0186359d12ee3dcf0e7cd46a464cf05568b1c2eb0ca7556e" Dec 01 08:47:02 crc kubenswrapper[4744]: I1201 08:47:02.837563 4744 scope.go:117] "RemoveContainer" containerID="c1420364bf8ee1eb684aa6d79a949049e0a5003a9c0df563fb01f07b375bfb9f" Dec 01 08:47:02 crc kubenswrapper[4744]: I1201 08:47:02.876153 4744 scope.go:117] "RemoveContainer" containerID="c5449e1ced9108117cf357cc5cde6963d71935c25dddf4f9c40fdfbab1f0aedb" Dec 01 08:47:08 crc kubenswrapper[4744]: I1201 08:47:08.290854 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:47:08 crc kubenswrapper[4744]: E1201 08:47:08.291646 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:47:11 crc kubenswrapper[4744]: I1201 08:47:11.054798 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t86st"] Dec 01 08:47:11 crc kubenswrapper[4744]: I1201 08:47:11.070099 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-t86st"] Dec 01 08:47:12 crc kubenswrapper[4744]: I1201 08:47:12.297980 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522db2e5-3aed-45b0-a228-ab2b875a3cee" path="/var/lib/kubelet/pods/522db2e5-3aed-45b0-a228-ab2b875a3cee/volumes" Dec 01 08:47:19 crc kubenswrapper[4744]: I1201 08:47:19.285015 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:47:19 crc kubenswrapper[4744]: E1201 08:47:19.285809 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:47:30 crc kubenswrapper[4744]: I1201 08:47:30.285895 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:47:30 crc kubenswrapper[4744]: E1201 08:47:30.286803 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:47:33 crc kubenswrapper[4744]: I1201 08:47:33.051909 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-24cgv"] Dec 01 08:47:33 crc kubenswrapper[4744]: I1201 08:47:33.064495 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-jc998"] Dec 01 08:47:33 crc kubenswrapper[4744]: I1201 08:47:33.071197 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-24cgv"] Dec 01 08:47:33 crc kubenswrapper[4744]: I1201 08:47:33.087260 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-jc998"] Dec 01 08:47:34 crc kubenswrapper[4744]: I1201 08:47:34.294147 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf" path="/var/lib/kubelet/pods/040ceb2d-26d7-4ffa-a86d-7a6b09e69cdf/volumes" Dec 01 08:47:34 crc kubenswrapper[4744]: I1201 08:47:34.294695 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eb87b97-a941-413c-9778-29a593d44ef0" path="/var/lib/kubelet/pods/3eb87b97-a941-413c-9778-29a593d44ef0/volumes" Dec 01 08:47:35 crc kubenswrapper[4744]: I1201 08:47:35.347470 4744 generic.go:334] "Generic (PLEG): container finished" podID="b2c369ea-0c26-424a-af22-996d64ecf25d" containerID="ca4e14d9f6a46a57a7f41ec8c6e5ee035e997153a875e0a32a5c2c5c3ae4da6e" exitCode=0 Dec 01 08:47:35 crc kubenswrapper[4744]: I1201 08:47:35.347585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" event={"ID":"b2c369ea-0c26-424a-af22-996d64ecf25d","Type":"ContainerDied","Data":"ca4e14d9f6a46a57a7f41ec8c6e5ee035e997153a875e0a32a5c2c5c3ae4da6e"} Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.798129 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.886944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-inventory\") pod \"b2c369ea-0c26-424a-af22-996d64ecf25d\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.887050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wnsk\" (UniqueName: \"kubernetes.io/projected/b2c369ea-0c26-424a-af22-996d64ecf25d-kube-api-access-9wnsk\") pod \"b2c369ea-0c26-424a-af22-996d64ecf25d\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.887106 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-ssh-key\") pod \"b2c369ea-0c26-424a-af22-996d64ecf25d\" (UID: \"b2c369ea-0c26-424a-af22-996d64ecf25d\") " Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.891970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2c369ea-0c26-424a-af22-996d64ecf25d-kube-api-access-9wnsk" (OuterVolumeSpecName: "kube-api-access-9wnsk") pod "b2c369ea-0c26-424a-af22-996d64ecf25d" (UID: "b2c369ea-0c26-424a-af22-996d64ecf25d"). InnerVolumeSpecName "kube-api-access-9wnsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.921166 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b2c369ea-0c26-424a-af22-996d64ecf25d" (UID: "b2c369ea-0c26-424a-af22-996d64ecf25d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.925024 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-inventory" (OuterVolumeSpecName: "inventory") pod "b2c369ea-0c26-424a-af22-996d64ecf25d" (UID: "b2c369ea-0c26-424a-af22-996d64ecf25d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.989014 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.989047 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wnsk\" (UniqueName: \"kubernetes.io/projected/b2c369ea-0c26-424a-af22-996d64ecf25d-kube-api-access-9wnsk\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:36 crc kubenswrapper[4744]: I1201 08:47:36.989059 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2c369ea-0c26-424a-af22-996d64ecf25d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.366868 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" event={"ID":"b2c369ea-0c26-424a-af22-996d64ecf25d","Type":"ContainerDied","Data":"545b89f8c73334fa8ad097918981dc0ba085f31b376889abe09deae528588b7b"} Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.367114 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="545b89f8c73334fa8ad097918981dc0ba085f31b376889abe09deae528588b7b" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.366906 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fttwk" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.462159 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f"] Dec 01 08:47:37 crc kubenswrapper[4744]: E1201 08:47:37.462649 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c369ea-0c26-424a-af22-996d64ecf25d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.462673 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c369ea-0c26-424a-af22-996d64ecf25d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.462953 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2c369ea-0c26-424a-af22-996d64ecf25d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.463749 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.467046 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.467940 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.468210 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.469078 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.478194 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f"] Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.601972 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.602111 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.602498 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjtgw\" (UniqueName: \"kubernetes.io/projected/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-kube-api-access-tjtgw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.704325 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjtgw\" (UniqueName: \"kubernetes.io/projected/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-kube-api-access-tjtgw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.704458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.704540 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.710608 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.711129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.722472 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjtgw\" (UniqueName: \"kubernetes.io/projected/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-kube-api-access-tjtgw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:37 crc kubenswrapper[4744]: I1201 08:47:37.787275 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:47:38 crc kubenswrapper[4744]: I1201 08:47:38.357127 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f"] Dec 01 08:47:38 crc kubenswrapper[4744]: I1201 08:47:38.378373 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" event={"ID":"e325084e-1abd-4b89-99f7-ec9e4f6f55ef","Type":"ContainerStarted","Data":"7ee5169ca4423a2b1d56911f5d94c02bd990980ff7a86ce54451bc4651e3fc15"} Dec 01 08:47:39 crc kubenswrapper[4744]: I1201 08:47:39.387124 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" event={"ID":"e325084e-1abd-4b89-99f7-ec9e4f6f55ef","Type":"ContainerStarted","Data":"e3aaffaaa154d4d34dbe10a40d10e51acc2edb9f7c213aed1d91a189896326bb"} Dec 01 08:47:44 crc kubenswrapper[4744]: I1201 08:47:44.285444 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:47:44 crc kubenswrapper[4744]: E1201 08:47:44.288317 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:47:59 crc kubenswrapper[4744]: I1201 08:47:59.285386 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:47:59 crc kubenswrapper[4744]: E1201 08:47:59.286387 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:48:03 crc kubenswrapper[4744]: I1201 08:48:03.012282 4744 scope.go:117] "RemoveContainer" containerID="b2a2a496c6333eb838de95bcb18cdb525d89b52ecaaeee3af278068c84bf73bf" Dec 01 08:48:03 crc kubenswrapper[4744]: I1201 08:48:03.089218 4744 scope.go:117] "RemoveContainer" containerID="789cde6612119cbc9df1a36a57e6970528223e416fbfee18e1cb1cfe62d93173" Dec 01 08:48:03 crc kubenswrapper[4744]: I1201 08:48:03.145081 4744 scope.go:117] "RemoveContainer" containerID="f12d78059710426c963a1cb1d2f3860cc3f000b0f18373f3a8687e7477a5cd6d" Dec 01 08:48:13 crc kubenswrapper[4744]: I1201 08:48:13.285337 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:48:13 crc kubenswrapper[4744]: E1201 08:48:13.286616 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:48:18 crc kubenswrapper[4744]: I1201 08:48:18.062713 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" podStartSLOduration=40.566397553 podStartE2EDuration="41.06268762s" podCreationTimestamp="2025-12-01 08:47:37 +0000 UTC" firstStartedPulling="2025-12-01 08:47:38.361527837 +0000 UTC m=+1790.350585778" lastFinishedPulling="2025-12-01 08:47:38.857817924 +0000 UTC m=+1790.846875845" observedRunningTime="2025-12-01 08:47:39.415074812 +0000 UTC m=+1791.404132733" watchObservedRunningTime="2025-12-01 08:48:18.06268762 +0000 UTC m=+1830.051745551" Dec 01 08:48:18 crc kubenswrapper[4744]: I1201 08:48:18.066852 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-klrlq"] Dec 01 08:48:18 crc kubenswrapper[4744]: I1201 08:48:18.076686 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-klrlq"] Dec 01 08:48:18 crc kubenswrapper[4744]: I1201 08:48:18.298510 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a49df21a-3b70-4797-8501-516a2f39e4f2" path="/var/lib/kubelet/pods/a49df21a-3b70-4797-8501-516a2f39e4f2/volumes" Dec 01 08:48:25 crc kubenswrapper[4744]: I1201 08:48:25.284776 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:48:25 crc kubenswrapper[4744]: E1201 08:48:25.285568 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:48:39 crc kubenswrapper[4744]: I1201 08:48:39.062838 4744 generic.go:334] "Generic (PLEG): container finished" podID="e325084e-1abd-4b89-99f7-ec9e4f6f55ef" containerID="e3aaffaaa154d4d34dbe10a40d10e51acc2edb9f7c213aed1d91a189896326bb" exitCode=0 Dec 01 08:48:39 crc kubenswrapper[4744]: I1201 08:48:39.062973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" event={"ID":"e325084e-1abd-4b89-99f7-ec9e4f6f55ef","Type":"ContainerDied","Data":"e3aaffaaa154d4d34dbe10a40d10e51acc2edb9f7c213aed1d91a189896326bb"} Dec 01 08:48:39 crc kubenswrapper[4744]: I1201 08:48:39.292545 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:48:39 crc kubenswrapper[4744]: E1201 08:48:39.292985 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.602621 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.722960 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjtgw\" (UniqueName: \"kubernetes.io/projected/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-kube-api-access-tjtgw\") pod \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.723012 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-ssh-key\") pod \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.723056 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-inventory\") pod \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\" (UID: \"e325084e-1abd-4b89-99f7-ec9e4f6f55ef\") " Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.737316 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-kube-api-access-tjtgw" (OuterVolumeSpecName: "kube-api-access-tjtgw") pod "e325084e-1abd-4b89-99f7-ec9e4f6f55ef" (UID: "e325084e-1abd-4b89-99f7-ec9e4f6f55ef"). InnerVolumeSpecName "kube-api-access-tjtgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.751258 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-inventory" (OuterVolumeSpecName: "inventory") pod "e325084e-1abd-4b89-99f7-ec9e4f6f55ef" (UID: "e325084e-1abd-4b89-99f7-ec9e4f6f55ef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.757056 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e325084e-1abd-4b89-99f7-ec9e4f6f55ef" (UID: "e325084e-1abd-4b89-99f7-ec9e4f6f55ef"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.826383 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjtgw\" (UniqueName: \"kubernetes.io/projected/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-kube-api-access-tjtgw\") on node \"crc\" DevicePath \"\"" Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.826434 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:48:40 crc kubenswrapper[4744]: I1201 08:48:40.826444 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e325084e-1abd-4b89-99f7-ec9e4f6f55ef-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.091066 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" event={"ID":"e325084e-1abd-4b89-99f7-ec9e4f6f55ef","Type":"ContainerDied","Data":"7ee5169ca4423a2b1d56911f5d94c02bd990980ff7a86ce54451bc4651e3fc15"} Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.091122 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ee5169ca4423a2b1d56911f5d94c02bd990980ff7a86ce54451bc4651e3fc15" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.091145 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.237092 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-g5snq"] Dec 01 08:48:41 crc kubenswrapper[4744]: E1201 08:48:41.237821 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e325084e-1abd-4b89-99f7-ec9e4f6f55ef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.237857 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e325084e-1abd-4b89-99f7-ec9e4f6f55ef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.238326 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e325084e-1abd-4b89-99f7-ec9e4f6f55ef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.239666 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.244376 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.244526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97cjh\" (UniqueName: \"kubernetes.io/projected/1e08d86e-f21d-42a3-8d93-be22013b97ac-kube-api-access-97cjh\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.244713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.249257 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-g5snq"] Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.267182 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.268985 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.269456 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.269555 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.346151 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.346599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.346717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97cjh\" (UniqueName: \"kubernetes.io/projected/1e08d86e-f21d-42a3-8d93-be22013b97ac-kube-api-access-97cjh\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.354355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.354813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.367889 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97cjh\" (UniqueName: \"kubernetes.io/projected/1e08d86e-f21d-42a3-8d93-be22013b97ac-kube-api-access-97cjh\") pod \"ssh-known-hosts-edpm-deployment-g5snq\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:41 crc kubenswrapper[4744]: I1201 08:48:41.595825 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:42 crc kubenswrapper[4744]: I1201 08:48:42.004495 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-g5snq"] Dec 01 08:48:42 crc kubenswrapper[4744]: I1201 08:48:42.101568 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" event={"ID":"1e08d86e-f21d-42a3-8d93-be22013b97ac","Type":"ContainerStarted","Data":"4506c0e34e0036dfce14e48d4127891a5c732d2f280bda6f44831de6bc4a03a5"} Dec 01 08:48:43 crc kubenswrapper[4744]: I1201 08:48:43.116101 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" event={"ID":"1e08d86e-f21d-42a3-8d93-be22013b97ac","Type":"ContainerStarted","Data":"7b6031a1f0aaab0b6e75fa052cbb822913c08b502b77100acf5c45e088156356"} Dec 01 08:48:43 crc kubenswrapper[4744]: I1201 08:48:43.138515 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" podStartSLOduration=1.6342382290000002 podStartE2EDuration="2.138491774s" podCreationTimestamp="2025-12-01 08:48:41 +0000 UTC" firstStartedPulling="2025-12-01 08:48:42.005509051 +0000 UTC m=+1853.994566972" lastFinishedPulling="2025-12-01 08:48:42.509762576 +0000 UTC m=+1854.498820517" observedRunningTime="2025-12-01 08:48:43.13309778 +0000 UTC m=+1855.122155711" watchObservedRunningTime="2025-12-01 08:48:43.138491774 +0000 UTC m=+1855.127549715" Dec 01 08:48:51 crc kubenswrapper[4744]: I1201 08:48:51.211449 4744 generic.go:334] "Generic (PLEG): container finished" podID="1e08d86e-f21d-42a3-8d93-be22013b97ac" containerID="7b6031a1f0aaab0b6e75fa052cbb822913c08b502b77100acf5c45e088156356" exitCode=0 Dec 01 08:48:51 crc kubenswrapper[4744]: I1201 08:48:51.212129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" event={"ID":"1e08d86e-f21d-42a3-8d93-be22013b97ac","Type":"ContainerDied","Data":"7b6031a1f0aaab0b6e75fa052cbb822913c08b502b77100acf5c45e088156356"} Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.672576 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.794954 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-ssh-key-openstack-edpm-ipam\") pod \"1e08d86e-f21d-42a3-8d93-be22013b97ac\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.795046 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97cjh\" (UniqueName: \"kubernetes.io/projected/1e08d86e-f21d-42a3-8d93-be22013b97ac-kube-api-access-97cjh\") pod \"1e08d86e-f21d-42a3-8d93-be22013b97ac\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.795156 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-inventory-0\") pod \"1e08d86e-f21d-42a3-8d93-be22013b97ac\" (UID: \"1e08d86e-f21d-42a3-8d93-be22013b97ac\") " Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.806258 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e08d86e-f21d-42a3-8d93-be22013b97ac-kube-api-access-97cjh" (OuterVolumeSpecName: "kube-api-access-97cjh") pod "1e08d86e-f21d-42a3-8d93-be22013b97ac" (UID: "1e08d86e-f21d-42a3-8d93-be22013b97ac"). InnerVolumeSpecName "kube-api-access-97cjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.822955 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1e08d86e-f21d-42a3-8d93-be22013b97ac" (UID: "1e08d86e-f21d-42a3-8d93-be22013b97ac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.823436 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "1e08d86e-f21d-42a3-8d93-be22013b97ac" (UID: "1e08d86e-f21d-42a3-8d93-be22013b97ac"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.898207 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.898263 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97cjh\" (UniqueName: \"kubernetes.io/projected/1e08d86e-f21d-42a3-8d93-be22013b97ac-kube-api-access-97cjh\") on node \"crc\" DevicePath \"\"" Dec 01 08:48:52 crc kubenswrapper[4744]: I1201 08:48:52.898285 4744 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1e08d86e-f21d-42a3-8d93-be22013b97ac-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.234154 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" event={"ID":"1e08d86e-f21d-42a3-8d93-be22013b97ac","Type":"ContainerDied","Data":"4506c0e34e0036dfce14e48d4127891a5c732d2f280bda6f44831de6bc4a03a5"} Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.234219 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4506c0e34e0036dfce14e48d4127891a5c732d2f280bda6f44831de6bc4a03a5" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.234298 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g5snq" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.360741 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz"] Dec 01 08:48:53 crc kubenswrapper[4744]: E1201 08:48:53.361606 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e08d86e-f21d-42a3-8d93-be22013b97ac" containerName="ssh-known-hosts-edpm-deployment" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.361647 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e08d86e-f21d-42a3-8d93-be22013b97ac" containerName="ssh-known-hosts-edpm-deployment" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.362024 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e08d86e-f21d-42a3-8d93-be22013b97ac" containerName="ssh-known-hosts-edpm-deployment" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.363143 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.365754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.366283 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.366563 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.370674 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.383577 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz"] Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.407757 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.407834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.407904 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88hlk\" (UniqueName: \"kubernetes.io/projected/dbf61995-97c3-420f-a591-5867f9192102-kube-api-access-88hlk\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.510327 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.510430 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.510498 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88hlk\" (UniqueName: \"kubernetes.io/projected/dbf61995-97c3-420f-a591-5867f9192102-kube-api-access-88hlk\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.517066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.523727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.529118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88hlk\" (UniqueName: \"kubernetes.io/projected/dbf61995-97c3-420f-a591-5867f9192102-kube-api-access-88hlk\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25vvz\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:53 crc kubenswrapper[4744]: I1201 08:48:53.699683 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:48:54 crc kubenswrapper[4744]: I1201 08:48:54.285801 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:48:54 crc kubenswrapper[4744]: E1201 08:48:54.286517 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:48:54 crc kubenswrapper[4744]: I1201 08:48:54.358078 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz"] Dec 01 08:48:55 crc kubenswrapper[4744]: I1201 08:48:55.258433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" event={"ID":"dbf61995-97c3-420f-a591-5867f9192102","Type":"ContainerStarted","Data":"181786c4c93c66febbccd1d6ac4867772221d666ead37ab73bbefb256994d3ed"} Dec 01 08:48:56 crc kubenswrapper[4744]: I1201 08:48:56.277077 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" event={"ID":"dbf61995-97c3-420f-a591-5867f9192102","Type":"ContainerStarted","Data":"5b61e991ec7a9ee126b54e22a3542c7a971caf308f9ed0ce176e88069598347c"} Dec 01 08:48:56 crc kubenswrapper[4744]: I1201 08:48:56.314058 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" podStartSLOduration=2.57849311 podStartE2EDuration="3.314028297s" podCreationTimestamp="2025-12-01 08:48:53 +0000 UTC" firstStartedPulling="2025-12-01 08:48:54.357400663 +0000 UTC m=+1866.346458584" lastFinishedPulling="2025-12-01 08:48:55.09293581 +0000 UTC m=+1867.081993771" observedRunningTime="2025-12-01 08:48:56.295662122 +0000 UTC m=+1868.284720083" watchObservedRunningTime="2025-12-01 08:48:56.314028297 +0000 UTC m=+1868.303086258" Dec 01 08:49:03 crc kubenswrapper[4744]: I1201 08:49:03.259441 4744 scope.go:117] "RemoveContainer" containerID="ec9a69156d333c2d20091a2617c10676d492dbcd7b26f3315c419be270edbd4c" Dec 01 08:49:04 crc kubenswrapper[4744]: I1201 08:49:04.392898 4744 generic.go:334] "Generic (PLEG): container finished" podID="dbf61995-97c3-420f-a591-5867f9192102" containerID="5b61e991ec7a9ee126b54e22a3542c7a971caf308f9ed0ce176e88069598347c" exitCode=0 Dec 01 08:49:04 crc kubenswrapper[4744]: I1201 08:49:04.393022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" event={"ID":"dbf61995-97c3-420f-a591-5867f9192102","Type":"ContainerDied","Data":"5b61e991ec7a9ee126b54e22a3542c7a971caf308f9ed0ce176e88069598347c"} Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.796707 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.888605 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88hlk\" (UniqueName: \"kubernetes.io/projected/dbf61995-97c3-420f-a591-5867f9192102-kube-api-access-88hlk\") pod \"dbf61995-97c3-420f-a591-5867f9192102\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.888700 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-inventory\") pod \"dbf61995-97c3-420f-a591-5867f9192102\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.888765 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-ssh-key\") pod \"dbf61995-97c3-420f-a591-5867f9192102\" (UID: \"dbf61995-97c3-420f-a591-5867f9192102\") " Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.895674 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbf61995-97c3-420f-a591-5867f9192102-kube-api-access-88hlk" (OuterVolumeSpecName: "kube-api-access-88hlk") pod "dbf61995-97c3-420f-a591-5867f9192102" (UID: "dbf61995-97c3-420f-a591-5867f9192102"). InnerVolumeSpecName "kube-api-access-88hlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.925420 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dbf61995-97c3-420f-a591-5867f9192102" (UID: "dbf61995-97c3-420f-a591-5867f9192102"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.927925 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-inventory" (OuterVolumeSpecName: "inventory") pod "dbf61995-97c3-420f-a591-5867f9192102" (UID: "dbf61995-97c3-420f-a591-5867f9192102"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.991352 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.991397 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbf61995-97c3-420f-a591-5867f9192102-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:05 crc kubenswrapper[4744]: I1201 08:49:05.991432 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88hlk\" (UniqueName: \"kubernetes.io/projected/dbf61995-97c3-420f-a591-5867f9192102-kube-api-access-88hlk\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.286275 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:49:06 crc kubenswrapper[4744]: E1201 08:49:06.286784 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.416845 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" event={"ID":"dbf61995-97c3-420f-a591-5867f9192102","Type":"ContainerDied","Data":"181786c4c93c66febbccd1d6ac4867772221d666ead37ab73bbefb256994d3ed"} Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.416883 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25vvz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.416909 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="181786c4c93c66febbccd1d6ac4867772221d666ead37ab73bbefb256994d3ed" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.519039 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz"] Dec 01 08:49:06 crc kubenswrapper[4744]: E1201 08:49:06.519489 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbf61995-97c3-420f-a591-5867f9192102" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.519509 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbf61995-97c3-420f-a591-5867f9192102" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.519753 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbf61995-97c3-420f-a591-5867f9192102" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.520744 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.523756 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.524238 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.524468 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.546748 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.554970 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz"] Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.604950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzbj4\" (UniqueName: \"kubernetes.io/projected/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-kube-api-access-lzbj4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.605117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.605271 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.706628 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.706666 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzbj4\" (UniqueName: \"kubernetes.io/projected/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-kube-api-access-lzbj4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.706739 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.729921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.730540 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.736931 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzbj4\" (UniqueName: \"kubernetes.io/projected/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-kube-api-access-lzbj4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:06 crc kubenswrapper[4744]: I1201 08:49:06.852640 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:07 crc kubenswrapper[4744]: I1201 08:49:07.240160 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz"] Dec 01 08:49:07 crc kubenswrapper[4744]: I1201 08:49:07.427632 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" event={"ID":"de4a3818-81e1-4735-b58d-6d1bb9df5ee2","Type":"ContainerStarted","Data":"64c4876a040859de44eb146462e629cc1a0507847f49c6c64de87ae7e4c0d495"} Dec 01 08:49:09 crc kubenswrapper[4744]: I1201 08:49:09.447967 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" event={"ID":"de4a3818-81e1-4735-b58d-6d1bb9df5ee2","Type":"ContainerStarted","Data":"04d5a902a0aceea4ef4a416fefef60a8811a139e9893b0a9dd74d594d97cfa62"} Dec 01 08:49:09 crc kubenswrapper[4744]: I1201 08:49:09.484343 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" podStartSLOduration=2.220192175 podStartE2EDuration="3.484300411s" podCreationTimestamp="2025-12-01 08:49:06 +0000 UTC" firstStartedPulling="2025-12-01 08:49:07.249909677 +0000 UTC m=+1879.238967598" lastFinishedPulling="2025-12-01 08:49:08.514017903 +0000 UTC m=+1880.503075834" observedRunningTime="2025-12-01 08:49:09.479507144 +0000 UTC m=+1881.468565105" watchObservedRunningTime="2025-12-01 08:49:09.484300411 +0000 UTC m=+1881.473358342" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.299315 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:49:18 crc kubenswrapper[4744]: E1201 08:49:18.300475 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.467149 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-twt6v"] Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.469137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.486598 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twt6v"] Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.521239 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-catalog-content\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.521320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht6cg\" (UniqueName: \"kubernetes.io/projected/3a335992-1166-433f-accd-2f94cca905ea-kube-api-access-ht6cg\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.521373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-utilities\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.624075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-catalog-content\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.624155 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht6cg\" (UniqueName: \"kubernetes.io/projected/3a335992-1166-433f-accd-2f94cca905ea-kube-api-access-ht6cg\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.624210 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-utilities\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.624674 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-catalog-content\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.624844 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-utilities\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.655399 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht6cg\" (UniqueName: \"kubernetes.io/projected/3a335992-1166-433f-accd-2f94cca905ea-kube-api-access-ht6cg\") pod \"community-operators-twt6v\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.666798 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vkwv8"] Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.668868 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.681936 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkwv8"] Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.789447 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.830209 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9p6f\" (UniqueName: \"kubernetes.io/projected/46c45775-cfd5-4f5c-a677-e06f7f8199ef-kube-api-access-d9p6f\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.830323 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-catalog-content\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.830436 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-utilities\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.957499 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9p6f\" (UniqueName: \"kubernetes.io/projected/46c45775-cfd5-4f5c-a677-e06f7f8199ef-kube-api-access-d9p6f\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.957980 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-catalog-content\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.958113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-utilities\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.958744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-utilities\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.960475 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-catalog-content\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:18 crc kubenswrapper[4744]: I1201 08:49:18.983593 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9p6f\" (UniqueName: \"kubernetes.io/projected/46c45775-cfd5-4f5c-a677-e06f7f8199ef-kube-api-access-d9p6f\") pod \"certified-operators-vkwv8\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.030014 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.158131 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twt6v"] Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.567182 4744 generic.go:334] "Generic (PLEG): container finished" podID="de4a3818-81e1-4735-b58d-6d1bb9df5ee2" containerID="04d5a902a0aceea4ef4a416fefef60a8811a139e9893b0a9dd74d594d97cfa62" exitCode=0 Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.567503 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" event={"ID":"de4a3818-81e1-4735-b58d-6d1bb9df5ee2","Type":"ContainerDied","Data":"04d5a902a0aceea4ef4a416fefef60a8811a139e9893b0a9dd74d594d97cfa62"} Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.570650 4744 generic.go:334] "Generic (PLEG): container finished" podID="3a335992-1166-433f-accd-2f94cca905ea" containerID="2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99" exitCode=0 Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.570676 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twt6v" event={"ID":"3a335992-1166-433f-accd-2f94cca905ea","Type":"ContainerDied","Data":"2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99"} Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.570691 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twt6v" event={"ID":"3a335992-1166-433f-accd-2f94cca905ea","Type":"ContainerStarted","Data":"38aafcc2fd418dc8adfe92aa7625e37a0f87a6c4e3aa2d50c31a8c74c7c2fdf1"} Dec 01 08:49:19 crc kubenswrapper[4744]: I1201 08:49:19.621038 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkwv8"] Dec 01 08:49:20 crc kubenswrapper[4744]: I1201 08:49:20.580281 4744 generic.go:334] "Generic (PLEG): container finished" podID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerID="2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903" exitCode=0 Dec 01 08:49:20 crc kubenswrapper[4744]: I1201 08:49:20.580361 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwv8" event={"ID":"46c45775-cfd5-4f5c-a677-e06f7f8199ef","Type":"ContainerDied","Data":"2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903"} Dec 01 08:49:20 crc kubenswrapper[4744]: I1201 08:49:20.580727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwv8" event={"ID":"46c45775-cfd5-4f5c-a677-e06f7f8199ef","Type":"ContainerStarted","Data":"333f7aa638e038e2544431d3018adc0cf303c92e35411250d4f39fa0394572e5"} Dec 01 08:49:20 crc kubenswrapper[4744]: I1201 08:49:20.985561 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.116246 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzbj4\" (UniqueName: \"kubernetes.io/projected/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-kube-api-access-lzbj4\") pod \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.116351 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-ssh-key\") pod \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.116428 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-inventory\") pod \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\" (UID: \"de4a3818-81e1-4735-b58d-6d1bb9df5ee2\") " Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.126747 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-kube-api-access-lzbj4" (OuterVolumeSpecName: "kube-api-access-lzbj4") pod "de4a3818-81e1-4735-b58d-6d1bb9df5ee2" (UID: "de4a3818-81e1-4735-b58d-6d1bb9df5ee2"). InnerVolumeSpecName "kube-api-access-lzbj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.142928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de4a3818-81e1-4735-b58d-6d1bb9df5ee2" (UID: "de4a3818-81e1-4735-b58d-6d1bb9df5ee2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.167647 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-inventory" (OuterVolumeSpecName: "inventory") pod "de4a3818-81e1-4735-b58d-6d1bb9df5ee2" (UID: "de4a3818-81e1-4735-b58d-6d1bb9df5ee2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.219454 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzbj4\" (UniqueName: \"kubernetes.io/projected/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-kube-api-access-lzbj4\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.219510 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.219530 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de4a3818-81e1-4735-b58d-6d1bb9df5ee2-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.593758 4744 generic.go:334] "Generic (PLEG): container finished" podID="3a335992-1166-433f-accd-2f94cca905ea" containerID="ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4" exitCode=0 Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.593897 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twt6v" event={"ID":"3a335992-1166-433f-accd-2f94cca905ea","Type":"ContainerDied","Data":"ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4"} Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.597814 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" event={"ID":"de4a3818-81e1-4735-b58d-6d1bb9df5ee2","Type":"ContainerDied","Data":"64c4876a040859de44eb146462e629cc1a0507847f49c6c64de87ae7e4c0d495"} Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.597850 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64c4876a040859de44eb146462e629cc1a0507847f49c6c64de87ae7e4c0d495" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.597887 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.709530 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7"] Dec 01 08:49:21 crc kubenswrapper[4744]: E1201 08:49:21.710007 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4a3818-81e1-4735-b58d-6d1bb9df5ee2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.710029 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4a3818-81e1-4735-b58d-6d1bb9df5ee2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.710256 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4a3818-81e1-4735-b58d-6d1bb9df5ee2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.711106 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.714549 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.716539 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.716687 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.716857 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.717034 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.717319 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.717357 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.718133 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.725478 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7"] Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838054 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838131 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkwmj\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-kube-api-access-wkwmj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838494 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838567 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838616 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838651 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838746 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.838797 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941496 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941678 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkwmj\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-kube-api-access-wkwmj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941835 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.941992 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.942046 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.942104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.942158 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.947761 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.948638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.950029 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.950394 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.951940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.952060 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.952067 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.952105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.952236 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.952768 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.954702 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.966335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.969059 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:21 crc kubenswrapper[4744]: I1201 08:49:21.979446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkwmj\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-kube-api-access-wkwmj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:22 crc kubenswrapper[4744]: I1201 08:49:22.039608 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:49:22 crc kubenswrapper[4744]: I1201 08:49:22.425737 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7"] Dec 01 08:49:22 crc kubenswrapper[4744]: I1201 08:49:22.608898 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" event={"ID":"cbfdfe61-094b-48f5-a450-0fcac89b8e8a","Type":"ContainerStarted","Data":"f9790b06cfdd6d9dd1482fe22a0135203f8c156c73e3f9406d03f83cb6ebadc5"} Dec 01 08:49:22 crc kubenswrapper[4744]: I1201 08:49:22.611244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twt6v" event={"ID":"3a335992-1166-433f-accd-2f94cca905ea","Type":"ContainerStarted","Data":"68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7"} Dec 01 08:49:22 crc kubenswrapper[4744]: I1201 08:49:22.613842 4744 generic.go:334] "Generic (PLEG): container finished" podID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerID="ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239" exitCode=0 Dec 01 08:49:22 crc kubenswrapper[4744]: I1201 08:49:22.613877 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwv8" event={"ID":"46c45775-cfd5-4f5c-a677-e06f7f8199ef","Type":"ContainerDied","Data":"ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239"} Dec 01 08:49:22 crc kubenswrapper[4744]: I1201 08:49:22.649179 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-twt6v" podStartSLOduration=1.880863365 podStartE2EDuration="4.64915809s" podCreationTimestamp="2025-12-01 08:49:18 +0000 UTC" firstStartedPulling="2025-12-01 08:49:19.572006328 +0000 UTC m=+1891.561064249" lastFinishedPulling="2025-12-01 08:49:22.340301043 +0000 UTC m=+1894.329358974" observedRunningTime="2025-12-01 08:49:22.638060623 +0000 UTC m=+1894.627118554" watchObservedRunningTime="2025-12-01 08:49:22.64915809 +0000 UTC m=+1894.638216011" Dec 01 08:49:23 crc kubenswrapper[4744]: I1201 08:49:23.626237 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwv8" event={"ID":"46c45775-cfd5-4f5c-a677-e06f7f8199ef","Type":"ContainerStarted","Data":"3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e"} Dec 01 08:49:23 crc kubenswrapper[4744]: I1201 08:49:23.628371 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" event={"ID":"cbfdfe61-094b-48f5-a450-0fcac89b8e8a","Type":"ContainerStarted","Data":"292104809c8ece82c4176ecaa1db0ec7bc71f3bad2d3da30a1bb7504a554f788"} Dec 01 08:49:23 crc kubenswrapper[4744]: I1201 08:49:23.647172 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vkwv8" podStartSLOduration=3.050902146 podStartE2EDuration="5.647154699s" podCreationTimestamp="2025-12-01 08:49:18 +0000 UTC" firstStartedPulling="2025-12-01 08:49:20.581951929 +0000 UTC m=+1892.571009860" lastFinishedPulling="2025-12-01 08:49:23.178204482 +0000 UTC m=+1895.167262413" observedRunningTime="2025-12-01 08:49:23.644118452 +0000 UTC m=+1895.633176393" watchObservedRunningTime="2025-12-01 08:49:23.647154699 +0000 UTC m=+1895.636212630" Dec 01 08:49:23 crc kubenswrapper[4744]: I1201 08:49:23.669690 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" podStartSLOduration=2.223848956 podStartE2EDuration="2.669669632s" podCreationTimestamp="2025-12-01 08:49:21 +0000 UTC" firstStartedPulling="2025-12-01 08:49:22.431203178 +0000 UTC m=+1894.420261109" lastFinishedPulling="2025-12-01 08:49:22.877023874 +0000 UTC m=+1894.866081785" observedRunningTime="2025-12-01 08:49:23.668661433 +0000 UTC m=+1895.657719354" watchObservedRunningTime="2025-12-01 08:49:23.669669632 +0000 UTC m=+1895.658727563" Dec 01 08:49:28 crc kubenswrapper[4744]: I1201 08:49:28.790125 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:28 crc kubenswrapper[4744]: I1201 08:49:28.790703 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:28 crc kubenswrapper[4744]: I1201 08:49:28.902558 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:29 crc kubenswrapper[4744]: I1201 08:49:29.031726 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:29 crc kubenswrapper[4744]: I1201 08:49:29.031764 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:29 crc kubenswrapper[4744]: I1201 08:49:29.105611 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:29 crc kubenswrapper[4744]: I1201 08:49:29.740954 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:29 crc kubenswrapper[4744]: I1201 08:49:29.765389 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:30 crc kubenswrapper[4744]: I1201 08:49:30.860967 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twt6v"] Dec 01 08:49:31 crc kubenswrapper[4744]: I1201 08:49:31.712848 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-twt6v" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="registry-server" containerID="cri-o://68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7" gracePeriod=2 Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.199116 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.270094 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ht6cg\" (UniqueName: \"kubernetes.io/projected/3a335992-1166-433f-accd-2f94cca905ea-kube-api-access-ht6cg\") pod \"3a335992-1166-433f-accd-2f94cca905ea\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.270537 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-catalog-content\") pod \"3a335992-1166-433f-accd-2f94cca905ea\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.270646 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-utilities\") pod \"3a335992-1166-433f-accd-2f94cca905ea\" (UID: \"3a335992-1166-433f-accd-2f94cca905ea\") " Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.274163 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-utilities" (OuterVolumeSpecName: "utilities") pod "3a335992-1166-433f-accd-2f94cca905ea" (UID: "3a335992-1166-433f-accd-2f94cca905ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.286707 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a335992-1166-433f-accd-2f94cca905ea-kube-api-access-ht6cg" (OuterVolumeSpecName: "kube-api-access-ht6cg") pod "3a335992-1166-433f-accd-2f94cca905ea" (UID: "3a335992-1166-433f-accd-2f94cca905ea"). InnerVolumeSpecName "kube-api-access-ht6cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.289772 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:49:32 crc kubenswrapper[4744]: E1201 08:49:32.289962 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.346554 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a335992-1166-433f-accd-2f94cca905ea" (UID: "3a335992-1166-433f-accd-2f94cca905ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.372943 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.373431 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a335992-1166-433f-accd-2f94cca905ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.373449 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ht6cg\" (UniqueName: \"kubernetes.io/projected/3a335992-1166-433f-accd-2f94cca905ea-kube-api-access-ht6cg\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.722369 4744 generic.go:334] "Generic (PLEG): container finished" podID="3a335992-1166-433f-accd-2f94cca905ea" containerID="68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7" exitCode=0 Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.722436 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twt6v" event={"ID":"3a335992-1166-433f-accd-2f94cca905ea","Type":"ContainerDied","Data":"68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7"} Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.722456 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twt6v" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.722485 4744 scope.go:117] "RemoveContainer" containerID="68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.722471 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twt6v" event={"ID":"3a335992-1166-433f-accd-2f94cca905ea","Type":"ContainerDied","Data":"38aafcc2fd418dc8adfe92aa7625e37a0f87a6c4e3aa2d50c31a8c74c7c2fdf1"} Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.766013 4744 scope.go:117] "RemoveContainer" containerID="ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.773306 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twt6v"] Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.782133 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-twt6v"] Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.803157 4744 scope.go:117] "RemoveContainer" containerID="2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.842268 4744 scope.go:117] "RemoveContainer" containerID="68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7" Dec 01 08:49:32 crc kubenswrapper[4744]: E1201 08:49:32.842744 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7\": container with ID starting with 68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7 not found: ID does not exist" containerID="68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.842808 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7"} err="failed to get container status \"68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7\": rpc error: code = NotFound desc = could not find container \"68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7\": container with ID starting with 68ba31eed39d8bb0c46d349e949880c27bad8b6d7203aef792926d7cbdfbc4e7 not found: ID does not exist" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.842849 4744 scope.go:117] "RemoveContainer" containerID="ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4" Dec 01 08:49:32 crc kubenswrapper[4744]: E1201 08:49:32.843287 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4\": container with ID starting with ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4 not found: ID does not exist" containerID="ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.843321 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4"} err="failed to get container status \"ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4\": rpc error: code = NotFound desc = could not find container \"ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4\": container with ID starting with ef9dc2b1b67dba367fff2f32372f4fac594b3058c05968b289de026cc6f89ea4 not found: ID does not exist" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.843344 4744 scope.go:117] "RemoveContainer" containerID="2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99" Dec 01 08:49:32 crc kubenswrapper[4744]: E1201 08:49:32.843642 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99\": container with ID starting with 2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99 not found: ID does not exist" containerID="2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99" Dec 01 08:49:32 crc kubenswrapper[4744]: I1201 08:49:32.843697 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99"} err="failed to get container status \"2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99\": rpc error: code = NotFound desc = could not find container \"2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99\": container with ID starting with 2bbf03455f3b445fb3d18e32f1d89dd8ec1e52f06738bf5506e6589010883c99 not found: ID does not exist" Dec 01 08:49:34 crc kubenswrapper[4744]: I1201 08:49:34.301887 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a335992-1166-433f-accd-2f94cca905ea" path="/var/lib/kubelet/pods/3a335992-1166-433f-accd-2f94cca905ea/volumes" Dec 01 08:49:35 crc kubenswrapper[4744]: I1201 08:49:35.653239 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkwv8"] Dec 01 08:49:35 crc kubenswrapper[4744]: I1201 08:49:35.653763 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vkwv8" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="registry-server" containerID="cri-o://3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e" gracePeriod=2 Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.112066 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.150384 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-catalog-content\") pod \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.150680 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9p6f\" (UniqueName: \"kubernetes.io/projected/46c45775-cfd5-4f5c-a677-e06f7f8199ef-kube-api-access-d9p6f\") pod \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.150735 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-utilities\") pod \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\" (UID: \"46c45775-cfd5-4f5c-a677-e06f7f8199ef\") " Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.151446 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-utilities" (OuterVolumeSpecName: "utilities") pod "46c45775-cfd5-4f5c-a677-e06f7f8199ef" (UID: "46c45775-cfd5-4f5c-a677-e06f7f8199ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.160458 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46c45775-cfd5-4f5c-a677-e06f7f8199ef-kube-api-access-d9p6f" (OuterVolumeSpecName: "kube-api-access-d9p6f") pod "46c45775-cfd5-4f5c-a677-e06f7f8199ef" (UID: "46c45775-cfd5-4f5c-a677-e06f7f8199ef"). InnerVolumeSpecName "kube-api-access-d9p6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.196840 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46c45775-cfd5-4f5c-a677-e06f7f8199ef" (UID: "46c45775-cfd5-4f5c-a677-e06f7f8199ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.252817 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9p6f\" (UniqueName: \"kubernetes.io/projected/46c45775-cfd5-4f5c-a677-e06f7f8199ef-kube-api-access-d9p6f\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.252851 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.252860 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c45775-cfd5-4f5c-a677-e06f7f8199ef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.777251 4744 generic.go:334] "Generic (PLEG): container finished" podID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerID="3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e" exitCode=0 Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.777292 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwv8" event={"ID":"46c45775-cfd5-4f5c-a677-e06f7f8199ef","Type":"ContainerDied","Data":"3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e"} Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.777322 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwv8" event={"ID":"46c45775-cfd5-4f5c-a677-e06f7f8199ef","Type":"ContainerDied","Data":"333f7aa638e038e2544431d3018adc0cf303c92e35411250d4f39fa0394572e5"} Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.777341 4744 scope.go:117] "RemoveContainer" containerID="3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.777497 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwv8" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.822136 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkwv8"] Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.830029 4744 scope.go:117] "RemoveContainer" containerID="ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.832748 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vkwv8"] Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.856608 4744 scope.go:117] "RemoveContainer" containerID="2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.907182 4744 scope.go:117] "RemoveContainer" containerID="3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e" Dec 01 08:49:36 crc kubenswrapper[4744]: E1201 08:49:36.917003 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e\": container with ID starting with 3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e not found: ID does not exist" containerID="3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.917043 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e"} err="failed to get container status \"3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e\": rpc error: code = NotFound desc = could not find container \"3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e\": container with ID starting with 3b1cac3e4333914831b1acf44ebcbf6825f510c00c100514e45f93cc3fc8f95e not found: ID does not exist" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.917069 4744 scope.go:117] "RemoveContainer" containerID="ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239" Dec 01 08:49:36 crc kubenswrapper[4744]: E1201 08:49:36.917607 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239\": container with ID starting with ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239 not found: ID does not exist" containerID="ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.917625 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239"} err="failed to get container status \"ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239\": rpc error: code = NotFound desc = could not find container \"ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239\": container with ID starting with ff59366734823990eeb1b7694a5bfb9585b84e9d2213aeb2cee4c8372c3bb239 not found: ID does not exist" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.917639 4744 scope.go:117] "RemoveContainer" containerID="2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903" Dec 01 08:49:36 crc kubenswrapper[4744]: E1201 08:49:36.918041 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903\": container with ID starting with 2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903 not found: ID does not exist" containerID="2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903" Dec 01 08:49:36 crc kubenswrapper[4744]: I1201 08:49:36.918061 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903"} err="failed to get container status \"2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903\": rpc error: code = NotFound desc = could not find container \"2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903\": container with ID starting with 2b341c46b3e78b08ff870bf7a66e39dd20de76fadb2b5166eb6df4b9c05c8903 not found: ID does not exist" Dec 01 08:49:38 crc kubenswrapper[4744]: I1201 08:49:38.297553 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" path="/var/lib/kubelet/pods/46c45775-cfd5-4f5c-a677-e06f7f8199ef/volumes" Dec 01 08:49:45 crc kubenswrapper[4744]: I1201 08:49:45.286991 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:49:45 crc kubenswrapper[4744]: E1201 08:49:45.288021 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:49:59 crc kubenswrapper[4744]: I1201 08:49:59.285318 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:50:00 crc kubenswrapper[4744]: I1201 08:50:00.023076 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"f9cd10e5091b95dcacd71252a807bcb4d070f147494bdbc17fae2776bb2df7b3"} Dec 01 08:50:11 crc kubenswrapper[4744]: I1201 08:50:11.143187 4744 generic.go:334] "Generic (PLEG): container finished" podID="cbfdfe61-094b-48f5-a450-0fcac89b8e8a" containerID="292104809c8ece82c4176ecaa1db0ec7bc71f3bad2d3da30a1bb7504a554f788" exitCode=0 Dec 01 08:50:11 crc kubenswrapper[4744]: I1201 08:50:11.143278 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" event={"ID":"cbfdfe61-094b-48f5-a450-0fcac89b8e8a","Type":"ContainerDied","Data":"292104809c8ece82c4176ecaa1db0ec7bc71f3bad2d3da30a1bb7504a554f788"} Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.662598 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.775714 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.775767 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-nova-combined-ca-bundle\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.775797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-inventory\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.775838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-libvirt-combined-ca-bundle\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.775901 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ovn-combined-ca-bundle\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.775958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-telemetry-combined-ca-bundle\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776039 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-neutron-metadata-combined-ca-bundle\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776089 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ssh-key\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776114 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkwmj\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-kube-api-access-wkwmj\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776211 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-repo-setup-combined-ca-bundle\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776230 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.776251 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-bootstrap-combined-ca-bundle\") pod \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\" (UID: \"cbfdfe61-094b-48f5-a450-0fcac89b8e8a\") " Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.783728 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.784508 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.784627 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.785291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.786007 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.786059 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.786434 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.787829 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-kube-api-access-wkwmj" (OuterVolumeSpecName: "kube-api-access-wkwmj") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "kube-api-access-wkwmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.788551 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.789215 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.791284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.795856 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.824628 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-inventory" (OuterVolumeSpecName: "inventory") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.825264 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cbfdfe61-094b-48f5-a450-0fcac89b8e8a" (UID: "cbfdfe61-094b-48f5-a450-0fcac89b8e8a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.877922 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878024 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878097 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878168 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878248 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkwmj\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-kube-api-access-wkwmj\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878330 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878427 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878497 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878552 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878609 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878684 4744 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878770 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878824 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:12 crc kubenswrapper[4744]: I1201 08:50:12.878885 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfdfe61-094b-48f5-a450-0fcac89b8e8a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.163661 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" event={"ID":"cbfdfe61-094b-48f5-a450-0fcac89b8e8a","Type":"ContainerDied","Data":"f9790b06cfdd6d9dd1482fe22a0135203f8c156c73e3f9406d03f83cb6ebadc5"} Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.164028 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9790b06cfdd6d9dd1482fe22a0135203f8c156c73e3f9406d03f83cb6ebadc5" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.163782 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.298710 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7"] Dec 01 08:50:13 crc kubenswrapper[4744]: E1201 08:50:13.299170 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbfdfe61-094b-48f5-a450-0fcac89b8e8a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299194 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbfdfe61-094b-48f5-a450-0fcac89b8e8a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 08:50:13 crc kubenswrapper[4744]: E1201 08:50:13.299211 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="extract-utilities" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299222 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="extract-utilities" Dec 01 08:50:13 crc kubenswrapper[4744]: E1201 08:50:13.299239 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="extract-content" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299247 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="extract-content" Dec 01 08:50:13 crc kubenswrapper[4744]: E1201 08:50:13.299264 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="extract-content" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299272 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="extract-content" Dec 01 08:50:13 crc kubenswrapper[4744]: E1201 08:50:13.299289 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="registry-server" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299297 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="registry-server" Dec 01 08:50:13 crc kubenswrapper[4744]: E1201 08:50:13.299322 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="extract-utilities" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299331 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="extract-utilities" Dec 01 08:50:13 crc kubenswrapper[4744]: E1201 08:50:13.299345 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="registry-server" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299353 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="registry-server" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299589 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="46c45775-cfd5-4f5c-a677-e06f7f8199ef" containerName="registry-server" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299615 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbfdfe61-094b-48f5-a450-0fcac89b8e8a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.299633 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a335992-1166-433f-accd-2f94cca905ea" containerName="registry-server" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.300368 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.304898 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.304982 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.305129 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.306853 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.307958 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.325911 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7"] Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.488468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.488696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.488854 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8dhd\" (UniqueName: \"kubernetes.io/projected/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-kube-api-access-b8dhd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.489131 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.489299 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.590518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.590600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.590646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8dhd\" (UniqueName: \"kubernetes.io/projected/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-kube-api-access-b8dhd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.590719 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.590771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.592222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.598129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.598367 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.598809 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.614282 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8dhd\" (UniqueName: \"kubernetes.io/projected/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-kube-api-access-b8dhd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7v6p7\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:13 crc kubenswrapper[4744]: I1201 08:50:13.625062 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:50:14 crc kubenswrapper[4744]: I1201 08:50:14.171758 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7"] Dec 01 08:50:14 crc kubenswrapper[4744]: W1201 08:50:14.179524 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25afa440_2166_4a7f_a6b9_8d4dc2c7087c.slice/crio-686e8cf7d92e48ad8b47bc36d5c0a9b718544c0695fd25b7830337ff9b977e2a WatchSource:0}: Error finding container 686e8cf7d92e48ad8b47bc36d5c0a9b718544c0695fd25b7830337ff9b977e2a: Status 404 returned error can't find the container with id 686e8cf7d92e48ad8b47bc36d5c0a9b718544c0695fd25b7830337ff9b977e2a Dec 01 08:50:14 crc kubenswrapper[4744]: I1201 08:50:14.183428 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:50:15 crc kubenswrapper[4744]: I1201 08:50:15.205084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" event={"ID":"25afa440-2166-4a7f-a6b9-8d4dc2c7087c","Type":"ContainerStarted","Data":"e955ebbf9d8c250ec06809697c2f07feb4a0da1e7eba5eb1e2ffcb811c0dc9c2"} Dec 01 08:50:15 crc kubenswrapper[4744]: I1201 08:50:15.206940 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" event={"ID":"25afa440-2166-4a7f-a6b9-8d4dc2c7087c","Type":"ContainerStarted","Data":"686e8cf7d92e48ad8b47bc36d5c0a9b718544c0695fd25b7830337ff9b977e2a"} Dec 01 08:50:15 crc kubenswrapper[4744]: I1201 08:50:15.230023 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" podStartSLOduration=1.500154252 podStartE2EDuration="2.230007835s" podCreationTimestamp="2025-12-01 08:50:13 +0000 UTC" firstStartedPulling="2025-12-01 08:50:14.183231742 +0000 UTC m=+1946.172289663" lastFinishedPulling="2025-12-01 08:50:14.913085295 +0000 UTC m=+1946.902143246" observedRunningTime="2025-12-01 08:50:15.220980783 +0000 UTC m=+1947.210038714" watchObservedRunningTime="2025-12-01 08:50:15.230007835 +0000 UTC m=+1947.219065756" Dec 01 08:51:29 crc kubenswrapper[4744]: I1201 08:51:29.007021 4744 generic.go:334] "Generic (PLEG): container finished" podID="25afa440-2166-4a7f-a6b9-8d4dc2c7087c" containerID="e955ebbf9d8c250ec06809697c2f07feb4a0da1e7eba5eb1e2ffcb811c0dc9c2" exitCode=0 Dec 01 08:51:29 crc kubenswrapper[4744]: I1201 08:51:29.007109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" event={"ID":"25afa440-2166-4a7f-a6b9-8d4dc2c7087c","Type":"ContainerDied","Data":"e955ebbf9d8c250ec06809697c2f07feb4a0da1e7eba5eb1e2ffcb811c0dc9c2"} Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.457481 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.563348 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovn-combined-ca-bundle\") pod \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.563555 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-inventory\") pod \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.563612 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovncontroller-config-0\") pod \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.563693 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ssh-key\") pod \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.563768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8dhd\" (UniqueName: \"kubernetes.io/projected/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-kube-api-access-b8dhd\") pod \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\" (UID: \"25afa440-2166-4a7f-a6b9-8d4dc2c7087c\") " Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.570789 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "25afa440-2166-4a7f-a6b9-8d4dc2c7087c" (UID: "25afa440-2166-4a7f-a6b9-8d4dc2c7087c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.570986 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-kube-api-access-b8dhd" (OuterVolumeSpecName: "kube-api-access-b8dhd") pod "25afa440-2166-4a7f-a6b9-8d4dc2c7087c" (UID: "25afa440-2166-4a7f-a6b9-8d4dc2c7087c"). InnerVolumeSpecName "kube-api-access-b8dhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.591494 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "25afa440-2166-4a7f-a6b9-8d4dc2c7087c" (UID: "25afa440-2166-4a7f-a6b9-8d4dc2c7087c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.596128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-inventory" (OuterVolumeSpecName: "inventory") pod "25afa440-2166-4a7f-a6b9-8d4dc2c7087c" (UID: "25afa440-2166-4a7f-a6b9-8d4dc2c7087c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.599430 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "25afa440-2166-4a7f-a6b9-8d4dc2c7087c" (UID: "25afa440-2166-4a7f-a6b9-8d4dc2c7087c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.666825 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8dhd\" (UniqueName: \"kubernetes.io/projected/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-kube-api-access-b8dhd\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.666847 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.666856 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.666866 4744 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:30 crc kubenswrapper[4744]: I1201 08:51:30.666874 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25afa440-2166-4a7f-a6b9-8d4dc2c7087c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.027191 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" event={"ID":"25afa440-2166-4a7f-a6b9-8d4dc2c7087c","Type":"ContainerDied","Data":"686e8cf7d92e48ad8b47bc36d5c0a9b718544c0695fd25b7830337ff9b977e2a"} Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.027251 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="686e8cf7d92e48ad8b47bc36d5c0a9b718544c0695fd25b7830337ff9b977e2a" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.027273 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7v6p7" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.193651 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp"] Dec 01 08:51:31 crc kubenswrapper[4744]: E1201 08:51:31.194055 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25afa440-2166-4a7f-a6b9-8d4dc2c7087c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.194076 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="25afa440-2166-4a7f-a6b9-8d4dc2c7087c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.194309 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="25afa440-2166-4a7f-a6b9-8d4dc2c7087c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.203075 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.210142 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.210320 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.210484 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.210596 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.210597 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.210765 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.246774 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp"] Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.279552 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn6jj\" (UniqueName: \"kubernetes.io/projected/ede60eff-c329-466b-a25d-7a309d424bf3-kube-api-access-fn6jj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.279604 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.279718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.279777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.279812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.279847 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.380879 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn6jj\" (UniqueName: \"kubernetes.io/projected/ede60eff-c329-466b-a25d-7a309d424bf3-kube-api-access-fn6jj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.381221 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.381291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.381338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.381376 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.381429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.385981 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.386404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.386466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.386890 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.389133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.402970 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn6jj\" (UniqueName: \"kubernetes.io/projected/ede60eff-c329-466b-a25d-7a309d424bf3-kube-api-access-fn6jj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:31 crc kubenswrapper[4744]: I1201 08:51:31.555025 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:51:32 crc kubenswrapper[4744]: I1201 08:51:32.145880 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp"] Dec 01 08:51:33 crc kubenswrapper[4744]: I1201 08:51:33.050204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" event={"ID":"ede60eff-c329-466b-a25d-7a309d424bf3","Type":"ContainerStarted","Data":"83606d177ad05df4e7f6df95d474d6827c68243c896226d8e1f34bdf79f172d2"} Dec 01 08:51:33 crc kubenswrapper[4744]: I1201 08:51:33.050665 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" event={"ID":"ede60eff-c329-466b-a25d-7a309d424bf3","Type":"ContainerStarted","Data":"54478121ef0d3230e6790d72cd1f76082eae3d88e952d4cae779d63a37b33163"} Dec 01 08:51:33 crc kubenswrapper[4744]: I1201 08:51:33.075609 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" podStartSLOduration=1.480466408 podStartE2EDuration="2.075585035s" podCreationTimestamp="2025-12-01 08:51:31 +0000 UTC" firstStartedPulling="2025-12-01 08:51:32.146211253 +0000 UTC m=+2024.135269164" lastFinishedPulling="2025-12-01 08:51:32.74132986 +0000 UTC m=+2024.730387791" observedRunningTime="2025-12-01 08:51:33.067666714 +0000 UTC m=+2025.056724655" watchObservedRunningTime="2025-12-01 08:51:33.075585035 +0000 UTC m=+2025.064642986" Dec 01 08:52:18 crc kubenswrapper[4744]: I1201 08:52:18.634535 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:52:18 crc kubenswrapper[4744]: I1201 08:52:18.635194 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:52:28 crc kubenswrapper[4744]: I1201 08:52:28.652816 4744 generic.go:334] "Generic (PLEG): container finished" podID="ede60eff-c329-466b-a25d-7a309d424bf3" containerID="83606d177ad05df4e7f6df95d474d6827c68243c896226d8e1f34bdf79f172d2" exitCode=0 Dec 01 08:52:28 crc kubenswrapper[4744]: I1201 08:52:28.652910 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" event={"ID":"ede60eff-c329-466b-a25d-7a309d424bf3","Type":"ContainerDied","Data":"83606d177ad05df4e7f6df95d474d6827c68243c896226d8e1f34bdf79f172d2"} Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.063629 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.162986 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-ssh-key\") pod \"ede60eff-c329-466b-a25d-7a309d424bf3\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.163361 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-nova-metadata-neutron-config-0\") pod \"ede60eff-c329-466b-a25d-7a309d424bf3\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.163521 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-metadata-combined-ca-bundle\") pod \"ede60eff-c329-466b-a25d-7a309d424bf3\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.163570 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ede60eff-c329-466b-a25d-7a309d424bf3\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.163619 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn6jj\" (UniqueName: \"kubernetes.io/projected/ede60eff-c329-466b-a25d-7a309d424bf3-kube-api-access-fn6jj\") pod \"ede60eff-c329-466b-a25d-7a309d424bf3\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.163684 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-inventory\") pod \"ede60eff-c329-466b-a25d-7a309d424bf3\" (UID: \"ede60eff-c329-466b-a25d-7a309d424bf3\") " Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.173733 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ede60eff-c329-466b-a25d-7a309d424bf3" (UID: "ede60eff-c329-466b-a25d-7a309d424bf3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.178710 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede60eff-c329-466b-a25d-7a309d424bf3-kube-api-access-fn6jj" (OuterVolumeSpecName: "kube-api-access-fn6jj") pod "ede60eff-c329-466b-a25d-7a309d424bf3" (UID: "ede60eff-c329-466b-a25d-7a309d424bf3"). InnerVolumeSpecName "kube-api-access-fn6jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.191875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-inventory" (OuterVolumeSpecName: "inventory") pod "ede60eff-c329-466b-a25d-7a309d424bf3" (UID: "ede60eff-c329-466b-a25d-7a309d424bf3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.192998 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ede60eff-c329-466b-a25d-7a309d424bf3" (UID: "ede60eff-c329-466b-a25d-7a309d424bf3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.198211 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ede60eff-c329-466b-a25d-7a309d424bf3" (UID: "ede60eff-c329-466b-a25d-7a309d424bf3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.208332 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ede60eff-c329-466b-a25d-7a309d424bf3" (UID: "ede60eff-c329-466b-a25d-7a309d424bf3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.266285 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.266326 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.266342 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn6jj\" (UniqueName: \"kubernetes.io/projected/ede60eff-c329-466b-a25d-7a309d424bf3-kube-api-access-fn6jj\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.266352 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.266360 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.266368 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ede60eff-c329-466b-a25d-7a309d424bf3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.677818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" event={"ID":"ede60eff-c329-466b-a25d-7a309d424bf3","Type":"ContainerDied","Data":"54478121ef0d3230e6790d72cd1f76082eae3d88e952d4cae779d63a37b33163"} Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.677921 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.677932 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54478121ef0d3230e6790d72cd1f76082eae3d88e952d4cae779d63a37b33163" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.800567 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm"] Dec 01 08:52:30 crc kubenswrapper[4744]: E1201 08:52:30.801185 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede60eff-c329-466b-a25d-7a309d424bf3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.801218 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede60eff-c329-466b-a25d-7a309d424bf3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.801568 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ede60eff-c329-466b-a25d-7a309d424bf3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.802357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.806875 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.806964 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.807249 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.807650 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.815368 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.818839 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm"] Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.980461 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.980781 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.980944 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79cw9\" (UniqueName: \"kubernetes.io/projected/b9ea1f1e-e881-4097-bd64-59ad7d48d552-kube-api-access-79cw9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.981046 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:30 crc kubenswrapper[4744]: I1201 08:52:30.981110 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.082558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.082623 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79cw9\" (UniqueName: \"kubernetes.io/projected/b9ea1f1e-e881-4097-bd64-59ad7d48d552-kube-api-access-79cw9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.082662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.082698 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.082774 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.088190 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.088838 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.090987 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.094963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.102207 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79cw9\" (UniqueName: \"kubernetes.io/projected/b9ea1f1e-e881-4097-bd64-59ad7d48d552-kube-api-access-79cw9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.125862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.656656 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm"] Dec 01 08:52:31 crc kubenswrapper[4744]: I1201 08:52:31.687427 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" event={"ID":"b9ea1f1e-e881-4097-bd64-59ad7d48d552","Type":"ContainerStarted","Data":"1b8adb8a2f7b7783819bb3a19150b381a9c78cb14f76636b97ef2a7a3d43e353"} Dec 01 08:52:32 crc kubenswrapper[4744]: I1201 08:52:32.699515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" event={"ID":"b9ea1f1e-e881-4097-bd64-59ad7d48d552","Type":"ContainerStarted","Data":"332290d3f8c86e53c4c7e63d95501c578c86cfac63967c975e27e5e0003110ea"} Dec 01 08:52:32 crc kubenswrapper[4744]: I1201 08:52:32.723185 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" podStartSLOduration=2.289537386 podStartE2EDuration="2.723162839s" podCreationTimestamp="2025-12-01 08:52:30 +0000 UTC" firstStartedPulling="2025-12-01 08:52:31.657374793 +0000 UTC m=+2083.646432754" lastFinishedPulling="2025-12-01 08:52:32.091000256 +0000 UTC m=+2084.080058207" observedRunningTime="2025-12-01 08:52:32.718588551 +0000 UTC m=+2084.707646492" watchObservedRunningTime="2025-12-01 08:52:32.723162839 +0000 UTC m=+2084.712220770" Dec 01 08:52:48 crc kubenswrapper[4744]: I1201 08:52:48.634145 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:52:48 crc kubenswrapper[4744]: I1201 08:52:48.635953 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:53:18 crc kubenswrapper[4744]: I1201 08:53:18.634758 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:53:18 crc kubenswrapper[4744]: I1201 08:53:18.635369 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:53:18 crc kubenswrapper[4744]: I1201 08:53:18.635467 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:53:18 crc kubenswrapper[4744]: I1201 08:53:18.636351 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9cd10e5091b95dcacd71252a807bcb4d070f147494bdbc17fae2776bb2df7b3"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:53:18 crc kubenswrapper[4744]: I1201 08:53:18.636444 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://f9cd10e5091b95dcacd71252a807bcb4d070f147494bdbc17fae2776bb2df7b3" gracePeriod=600 Dec 01 08:53:19 crc kubenswrapper[4744]: I1201 08:53:19.250747 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="f9cd10e5091b95dcacd71252a807bcb4d070f147494bdbc17fae2776bb2df7b3" exitCode=0 Dec 01 08:53:19 crc kubenswrapper[4744]: I1201 08:53:19.250782 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"f9cd10e5091b95dcacd71252a807bcb4d070f147494bdbc17fae2776bb2df7b3"} Dec 01 08:53:19 crc kubenswrapper[4744]: I1201 08:53:19.251363 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc"} Dec 01 08:53:19 crc kubenswrapper[4744]: I1201 08:53:19.251390 4744 scope.go:117] "RemoveContainer" containerID="8b6c001ca066983ac42e1a1ff43ea30da86d76423cb42365aaa635ff1b6e3abd" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.462365 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v7gfr"] Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.466326 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.473103 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v7gfr"] Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.531735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-catalog-content\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.534818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msbcz\" (UniqueName: \"kubernetes.io/projected/3f4bcf0b-b977-4246-a079-9672cb61bdd8-kube-api-access-msbcz\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.535053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-utilities\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.637197 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-catalog-content\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.637297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msbcz\" (UniqueName: \"kubernetes.io/projected/3f4bcf0b-b977-4246-a079-9672cb61bdd8-kube-api-access-msbcz\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.637366 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-utilities\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.637769 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-catalog-content\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.637797 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-utilities\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.656911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msbcz\" (UniqueName: \"kubernetes.io/projected/3f4bcf0b-b977-4246-a079-9672cb61bdd8-kube-api-access-msbcz\") pod \"redhat-operators-v7gfr\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:29 crc kubenswrapper[4744]: I1201 08:54:29.806896 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:30 crc kubenswrapper[4744]: I1201 08:54:30.277708 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v7gfr"] Dec 01 08:54:31 crc kubenswrapper[4744]: I1201 08:54:31.094304 4744 generic.go:334] "Generic (PLEG): container finished" podID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerID="9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8" exitCode=0 Dec 01 08:54:31 crc kubenswrapper[4744]: I1201 08:54:31.095297 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7gfr" event={"ID":"3f4bcf0b-b977-4246-a079-9672cb61bdd8","Type":"ContainerDied","Data":"9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8"} Dec 01 08:54:31 crc kubenswrapper[4744]: I1201 08:54:31.095349 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7gfr" event={"ID":"3f4bcf0b-b977-4246-a079-9672cb61bdd8","Type":"ContainerStarted","Data":"8989b96077bec6d5f9cc348af9f21e88166f13f5eda1b2553cb583d576b57267"} Dec 01 08:54:32 crc kubenswrapper[4744]: I1201 08:54:32.110335 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7gfr" event={"ID":"3f4bcf0b-b977-4246-a079-9672cb61bdd8","Type":"ContainerStarted","Data":"78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73"} Dec 01 08:54:33 crc kubenswrapper[4744]: I1201 08:54:33.129165 4744 generic.go:334] "Generic (PLEG): container finished" podID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerID="78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73" exitCode=0 Dec 01 08:54:33 crc kubenswrapper[4744]: I1201 08:54:33.129400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7gfr" event={"ID":"3f4bcf0b-b977-4246-a079-9672cb61bdd8","Type":"ContainerDied","Data":"78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73"} Dec 01 08:54:35 crc kubenswrapper[4744]: I1201 08:54:35.150696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7gfr" event={"ID":"3f4bcf0b-b977-4246-a079-9672cb61bdd8","Type":"ContainerStarted","Data":"0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979"} Dec 01 08:54:35 crc kubenswrapper[4744]: I1201 08:54:35.167385 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v7gfr" podStartSLOduration=3.319778937 podStartE2EDuration="6.167364656s" podCreationTimestamp="2025-12-01 08:54:29 +0000 UTC" firstStartedPulling="2025-12-01 08:54:31.098975418 +0000 UTC m=+2203.088033379" lastFinishedPulling="2025-12-01 08:54:33.946561177 +0000 UTC m=+2205.935619098" observedRunningTime="2025-12-01 08:54:35.166711868 +0000 UTC m=+2207.155769829" watchObservedRunningTime="2025-12-01 08:54:35.167364656 +0000 UTC m=+2207.156422577" Dec 01 08:54:39 crc kubenswrapper[4744]: I1201 08:54:39.807631 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:39 crc kubenswrapper[4744]: I1201 08:54:39.808062 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:40 crc kubenswrapper[4744]: I1201 08:54:40.862827 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v7gfr" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="registry-server" probeResult="failure" output=< Dec 01 08:54:40 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 08:54:40 crc kubenswrapper[4744]: > Dec 01 08:54:49 crc kubenswrapper[4744]: I1201 08:54:49.903202 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:49 crc kubenswrapper[4744]: I1201 08:54:49.964898 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:50 crc kubenswrapper[4744]: I1201 08:54:50.159524 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v7gfr"] Dec 01 08:54:51 crc kubenswrapper[4744]: I1201 08:54:51.323671 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v7gfr" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="registry-server" containerID="cri-o://0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979" gracePeriod=2 Dec 01 08:54:51 crc kubenswrapper[4744]: I1201 08:54:51.814533 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:51 crc kubenswrapper[4744]: I1201 08:54:51.909272 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-catalog-content\") pod \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " Dec 01 08:54:51 crc kubenswrapper[4744]: I1201 08:54:51.909332 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msbcz\" (UniqueName: \"kubernetes.io/projected/3f4bcf0b-b977-4246-a079-9672cb61bdd8-kube-api-access-msbcz\") pod \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " Dec 01 08:54:51 crc kubenswrapper[4744]: I1201 08:54:51.909533 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-utilities\") pod \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\" (UID: \"3f4bcf0b-b977-4246-a079-9672cb61bdd8\") " Dec 01 08:54:51 crc kubenswrapper[4744]: I1201 08:54:51.910300 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-utilities" (OuterVolumeSpecName: "utilities") pod "3f4bcf0b-b977-4246-a079-9672cb61bdd8" (UID: "3f4bcf0b-b977-4246-a079-9672cb61bdd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:54:51 crc kubenswrapper[4744]: I1201 08:54:51.915124 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f4bcf0b-b977-4246-a079-9672cb61bdd8-kube-api-access-msbcz" (OuterVolumeSpecName: "kube-api-access-msbcz") pod "3f4bcf0b-b977-4246-a079-9672cb61bdd8" (UID: "3f4bcf0b-b977-4246-a079-9672cb61bdd8"). InnerVolumeSpecName "kube-api-access-msbcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.004975 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f4bcf0b-b977-4246-a079-9672cb61bdd8" (UID: "3f4bcf0b-b977-4246-a079-9672cb61bdd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.012266 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.012315 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msbcz\" (UniqueName: \"kubernetes.io/projected/3f4bcf0b-b977-4246-a079-9672cb61bdd8-kube-api-access-msbcz\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.012331 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4bcf0b-b977-4246-a079-9672cb61bdd8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.337260 4744 generic.go:334] "Generic (PLEG): container finished" podID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerID="0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979" exitCode=0 Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.337311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7gfr" event={"ID":"3f4bcf0b-b977-4246-a079-9672cb61bdd8","Type":"ContainerDied","Data":"0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979"} Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.337344 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7gfr" event={"ID":"3f4bcf0b-b977-4246-a079-9672cb61bdd8","Type":"ContainerDied","Data":"8989b96077bec6d5f9cc348af9f21e88166f13f5eda1b2553cb583d576b57267"} Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.337368 4744 scope.go:117] "RemoveContainer" containerID="0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.339040 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7gfr" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.369239 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v7gfr"] Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.378895 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v7gfr"] Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.390654 4744 scope.go:117] "RemoveContainer" containerID="78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.419568 4744 scope.go:117] "RemoveContainer" containerID="9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.476642 4744 scope.go:117] "RemoveContainer" containerID="0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979" Dec 01 08:54:52 crc kubenswrapper[4744]: E1201 08:54:52.477234 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979\": container with ID starting with 0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979 not found: ID does not exist" containerID="0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.477304 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979"} err="failed to get container status \"0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979\": rpc error: code = NotFound desc = could not find container \"0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979\": container with ID starting with 0140e1c77148d78bb04fa48ba55ec853bd0fed0e0f7da60bd5dd5fd2fef00979 not found: ID does not exist" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.477350 4744 scope.go:117] "RemoveContainer" containerID="78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73" Dec 01 08:54:52 crc kubenswrapper[4744]: E1201 08:54:52.478059 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73\": container with ID starting with 78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73 not found: ID does not exist" containerID="78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.478096 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73"} err="failed to get container status \"78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73\": rpc error: code = NotFound desc = could not find container \"78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73\": container with ID starting with 78441a126161b9e1a4db5fac456ebad5d8f9cf786acde443f350c7228372cf73 not found: ID does not exist" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.478114 4744 scope.go:117] "RemoveContainer" containerID="9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8" Dec 01 08:54:52 crc kubenswrapper[4744]: E1201 08:54:52.478452 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8\": container with ID starting with 9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8 not found: ID does not exist" containerID="9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8" Dec 01 08:54:52 crc kubenswrapper[4744]: I1201 08:54:52.478490 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8"} err="failed to get container status \"9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8\": rpc error: code = NotFound desc = could not find container \"9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8\": container with ID starting with 9321c217b8fa035c69d4327dc08dd078f16714c2857e0911ffe4b18f2f7937a8 not found: ID does not exist" Dec 01 08:54:54 crc kubenswrapper[4744]: I1201 08:54:54.303110 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" path="/var/lib/kubelet/pods/3f4bcf0b-b977-4246-a079-9672cb61bdd8/volumes" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.634444 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.634941 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.799840 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n7hwv"] Dec 01 08:55:18 crc kubenswrapper[4744]: E1201 08:55:18.800769 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="registry-server" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.800814 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="registry-server" Dec 01 08:55:18 crc kubenswrapper[4744]: E1201 08:55:18.800854 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="extract-content" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.800873 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="extract-content" Dec 01 08:55:18 crc kubenswrapper[4744]: E1201 08:55:18.800926 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="extract-utilities" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.800945 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="extract-utilities" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.801402 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f4bcf0b-b977-4246-a079-9672cb61bdd8" containerName="registry-server" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.804832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.835176 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7hwv"] Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.895892 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skhf9\" (UniqueName: \"kubernetes.io/projected/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-kube-api-access-skhf9\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.895992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-catalog-content\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.896337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-utilities\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.998022 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-utilities\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.998339 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skhf9\" (UniqueName: \"kubernetes.io/projected/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-kube-api-access-skhf9\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.998529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-utilities\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.998537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-catalog-content\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:18 crc kubenswrapper[4744]: I1201 08:55:18.998899 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-catalog-content\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:19 crc kubenswrapper[4744]: I1201 08:55:19.018178 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skhf9\" (UniqueName: \"kubernetes.io/projected/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-kube-api-access-skhf9\") pod \"redhat-marketplace-n7hwv\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:19 crc kubenswrapper[4744]: I1201 08:55:19.133451 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:19 crc kubenswrapper[4744]: I1201 08:55:19.578449 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7hwv"] Dec 01 08:55:19 crc kubenswrapper[4744]: W1201 08:55:19.588828 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1d5baf3_f21e_4ded_afa0_9f4c939bccb6.slice/crio-7ac9ebd561db9a549ceb009d524e8bac657d1cdba6fd78c2ed87080ab70f03df WatchSource:0}: Error finding container 7ac9ebd561db9a549ceb009d524e8bac657d1cdba6fd78c2ed87080ab70f03df: Status 404 returned error can't find the container with id 7ac9ebd561db9a549ceb009d524e8bac657d1cdba6fd78c2ed87080ab70f03df Dec 01 08:55:19 crc kubenswrapper[4744]: I1201 08:55:19.652880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7hwv" event={"ID":"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6","Type":"ContainerStarted","Data":"7ac9ebd561db9a549ceb009d524e8bac657d1cdba6fd78c2ed87080ab70f03df"} Dec 01 08:55:20 crc kubenswrapper[4744]: I1201 08:55:20.667485 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerID="ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863" exitCode=0 Dec 01 08:55:20 crc kubenswrapper[4744]: I1201 08:55:20.667610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7hwv" event={"ID":"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6","Type":"ContainerDied","Data":"ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863"} Dec 01 08:55:20 crc kubenswrapper[4744]: I1201 08:55:20.670837 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:55:22 crc kubenswrapper[4744]: I1201 08:55:22.693679 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerID="097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08" exitCode=0 Dec 01 08:55:22 crc kubenswrapper[4744]: I1201 08:55:22.693714 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7hwv" event={"ID":"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6","Type":"ContainerDied","Data":"097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08"} Dec 01 08:55:24 crc kubenswrapper[4744]: I1201 08:55:24.736278 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7hwv" event={"ID":"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6","Type":"ContainerStarted","Data":"794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c"} Dec 01 08:55:24 crc kubenswrapper[4744]: I1201 08:55:24.766568 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n7hwv" podStartSLOduration=3.953131126 podStartE2EDuration="6.766546416s" podCreationTimestamp="2025-12-01 08:55:18 +0000 UTC" firstStartedPulling="2025-12-01 08:55:20.670574162 +0000 UTC m=+2252.659632083" lastFinishedPulling="2025-12-01 08:55:23.483989432 +0000 UTC m=+2255.473047373" observedRunningTime="2025-12-01 08:55:24.760732354 +0000 UTC m=+2256.749790285" watchObservedRunningTime="2025-12-01 08:55:24.766546416 +0000 UTC m=+2256.755604337" Dec 01 08:55:29 crc kubenswrapper[4744]: I1201 08:55:29.134640 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:29 crc kubenswrapper[4744]: I1201 08:55:29.135390 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:29 crc kubenswrapper[4744]: I1201 08:55:29.202540 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:29 crc kubenswrapper[4744]: I1201 08:55:29.823967 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:29 crc kubenswrapper[4744]: I1201 08:55:29.871051 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7hwv"] Dec 01 08:55:31 crc kubenswrapper[4744]: I1201 08:55:31.802650 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n7hwv" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="registry-server" containerID="cri-o://794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c" gracePeriod=2 Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.306038 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.411628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-catalog-content\") pod \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.411689 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-utilities\") pod \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.411747 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skhf9\" (UniqueName: \"kubernetes.io/projected/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-kube-api-access-skhf9\") pod \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\" (UID: \"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6\") " Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.412944 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-utilities" (OuterVolumeSpecName: "utilities") pod "d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" (UID: "d1d5baf3-f21e-4ded-afa0-9f4c939bccb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.417128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-kube-api-access-skhf9" (OuterVolumeSpecName: "kube-api-access-skhf9") pod "d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" (UID: "d1d5baf3-f21e-4ded-afa0-9f4c939bccb6"). InnerVolumeSpecName "kube-api-access-skhf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.438196 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" (UID: "d1d5baf3-f21e-4ded-afa0-9f4c939bccb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.513676 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.513714 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skhf9\" (UniqueName: \"kubernetes.io/projected/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-kube-api-access-skhf9\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.513728 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.818399 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerID="794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c" exitCode=0 Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.818477 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7hwv" event={"ID":"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6","Type":"ContainerDied","Data":"794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c"} Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.820717 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7hwv" event={"ID":"d1d5baf3-f21e-4ded-afa0-9f4c939bccb6","Type":"ContainerDied","Data":"7ac9ebd561db9a549ceb009d524e8bac657d1cdba6fd78c2ed87080ab70f03df"} Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.820768 4744 scope.go:117] "RemoveContainer" containerID="794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.818633 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7hwv" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.854546 4744 scope.go:117] "RemoveContainer" containerID="097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.882200 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7hwv"] Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.891535 4744 scope.go:117] "RemoveContainer" containerID="ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.896113 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7hwv"] Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.923009 4744 scope.go:117] "RemoveContainer" containerID="794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c" Dec 01 08:55:32 crc kubenswrapper[4744]: E1201 08:55:32.924982 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c\": container with ID starting with 794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c not found: ID does not exist" containerID="794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.925037 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c"} err="failed to get container status \"794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c\": rpc error: code = NotFound desc = could not find container \"794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c\": container with ID starting with 794911b7f1423f78e23a5b4a6ebb08c3f2ad40e09394ce6b352c95356eeaf16c not found: ID does not exist" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.925071 4744 scope.go:117] "RemoveContainer" containerID="097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08" Dec 01 08:55:32 crc kubenswrapper[4744]: E1201 08:55:32.925461 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08\": container with ID starting with 097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08 not found: ID does not exist" containerID="097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.925489 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08"} err="failed to get container status \"097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08\": rpc error: code = NotFound desc = could not find container \"097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08\": container with ID starting with 097e971ba178244afd49e4c23c4987828f88498f765c8e2f8248d34b3a34be08 not found: ID does not exist" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.925506 4744 scope.go:117] "RemoveContainer" containerID="ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863" Dec 01 08:55:32 crc kubenswrapper[4744]: E1201 08:55:32.925831 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863\": container with ID starting with ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863 not found: ID does not exist" containerID="ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863" Dec 01 08:55:32 crc kubenswrapper[4744]: I1201 08:55:32.925887 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863"} err="failed to get container status \"ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863\": rpc error: code = NotFound desc = could not find container \"ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863\": container with ID starting with ac02b03b8f4d8dd0ff3666ae133bde50ca820f2668515a16c6bca72b38447863 not found: ID does not exist" Dec 01 08:55:34 crc kubenswrapper[4744]: I1201 08:55:34.309207 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" path="/var/lib/kubelet/pods/d1d5baf3-f21e-4ded-afa0-9f4c939bccb6/volumes" Dec 01 08:55:48 crc kubenswrapper[4744]: I1201 08:55:48.634971 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:55:48 crc kubenswrapper[4744]: I1201 08:55:48.635671 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:56:18 crc kubenswrapper[4744]: I1201 08:56:18.635114 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:56:18 crc kubenswrapper[4744]: I1201 08:56:18.635818 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:56:18 crc kubenswrapper[4744]: I1201 08:56:18.635886 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 08:56:18 crc kubenswrapper[4744]: I1201 08:56:18.636923 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:56:18 crc kubenswrapper[4744]: I1201 08:56:18.636999 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" gracePeriod=600 Dec 01 08:56:18 crc kubenswrapper[4744]: E1201 08:56:18.761355 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:56:19 crc kubenswrapper[4744]: I1201 08:56:19.294533 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" exitCode=0 Dec 01 08:56:19 crc kubenswrapper[4744]: I1201 08:56:19.294588 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc"} Dec 01 08:56:19 crc kubenswrapper[4744]: I1201 08:56:19.294633 4744 scope.go:117] "RemoveContainer" containerID="f9cd10e5091b95dcacd71252a807bcb4d070f147494bdbc17fae2776bb2df7b3" Dec 01 08:56:19 crc kubenswrapper[4744]: I1201 08:56:19.295233 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:56:19 crc kubenswrapper[4744]: E1201 08:56:19.295598 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:56:32 crc kubenswrapper[4744]: I1201 08:56:32.285891 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:56:32 crc kubenswrapper[4744]: E1201 08:56:32.287011 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:56:47 crc kubenswrapper[4744]: I1201 08:56:47.286198 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:56:47 crc kubenswrapper[4744]: E1201 08:56:47.287650 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:57:02 crc kubenswrapper[4744]: I1201 08:57:02.285827 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:57:02 crc kubenswrapper[4744]: E1201 08:57:02.286962 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:57:12 crc kubenswrapper[4744]: I1201 08:57:12.848147 4744 generic.go:334] "Generic (PLEG): container finished" podID="b9ea1f1e-e881-4097-bd64-59ad7d48d552" containerID="332290d3f8c86e53c4c7e63d95501c578c86cfac63967c975e27e5e0003110ea" exitCode=0 Dec 01 08:57:12 crc kubenswrapper[4744]: I1201 08:57:12.848228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" event={"ID":"b9ea1f1e-e881-4097-bd64-59ad7d48d552","Type":"ContainerDied","Data":"332290d3f8c86e53c4c7e63d95501c578c86cfac63967c975e27e5e0003110ea"} Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.297195 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.459753 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79cw9\" (UniqueName: \"kubernetes.io/projected/b9ea1f1e-e881-4097-bd64-59ad7d48d552-kube-api-access-79cw9\") pod \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.460150 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-combined-ca-bundle\") pod \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.460186 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-secret-0\") pod \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.460213 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-inventory\") pod \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.460251 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-ssh-key\") pod \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\" (UID: \"b9ea1f1e-e881-4097-bd64-59ad7d48d552\") " Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.466000 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ea1f1e-e881-4097-bd64-59ad7d48d552-kube-api-access-79cw9" (OuterVolumeSpecName: "kube-api-access-79cw9") pod "b9ea1f1e-e881-4097-bd64-59ad7d48d552" (UID: "b9ea1f1e-e881-4097-bd64-59ad7d48d552"). InnerVolumeSpecName "kube-api-access-79cw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.470550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b9ea1f1e-e881-4097-bd64-59ad7d48d552" (UID: "b9ea1f1e-e881-4097-bd64-59ad7d48d552"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.486385 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b9ea1f1e-e881-4097-bd64-59ad7d48d552" (UID: "b9ea1f1e-e881-4097-bd64-59ad7d48d552"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.500547 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-inventory" (OuterVolumeSpecName: "inventory") pod "b9ea1f1e-e881-4097-bd64-59ad7d48d552" (UID: "b9ea1f1e-e881-4097-bd64-59ad7d48d552"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.507498 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "b9ea1f1e-e881-4097-bd64-59ad7d48d552" (UID: "b9ea1f1e-e881-4097-bd64-59ad7d48d552"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.562845 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.562889 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.562905 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.562918 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9ea1f1e-e881-4097-bd64-59ad7d48d552-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.562932 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79cw9\" (UniqueName: \"kubernetes.io/projected/b9ea1f1e-e881-4097-bd64-59ad7d48d552-kube-api-access-79cw9\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.869225 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" event={"ID":"b9ea1f1e-e881-4097-bd64-59ad7d48d552","Type":"ContainerDied","Data":"1b8adb8a2f7b7783819bb3a19150b381a9c78cb14f76636b97ef2a7a3d43e353"} Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.869275 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.869293 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b8adb8a2f7b7783819bb3a19150b381a9c78cb14f76636b97ef2a7a3d43e353" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.958717 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6"] Dec 01 08:57:14 crc kubenswrapper[4744]: E1201 08:57:14.959273 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="registry-server" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.959304 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="registry-server" Dec 01 08:57:14 crc kubenswrapper[4744]: E1201 08:57:14.959333 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="extract-utilities" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.959348 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="extract-utilities" Dec 01 08:57:14 crc kubenswrapper[4744]: E1201 08:57:14.959398 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ea1f1e-e881-4097-bd64-59ad7d48d552" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.959438 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ea1f1e-e881-4097-bd64-59ad7d48d552" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 08:57:14 crc kubenswrapper[4744]: E1201 08:57:14.959460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="extract-content" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.959471 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="extract-content" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.959750 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d5baf3-f21e-4ded-afa0-9f4c939bccb6" containerName="registry-server" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.959802 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ea1f1e-e881-4097-bd64-59ad7d48d552" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.960556 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.963153 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.963230 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.963473 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.965451 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.965544 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.965547 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.966191 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 08:57:14 crc kubenswrapper[4744]: I1201 08:57:14.969521 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6"] Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070710 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hd5r\" (UniqueName: \"kubernetes.io/projected/da76f156-fbfc-46fe-b11d-a71a0bfd280d-kube-api-access-7hd5r\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070832 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070854 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070900 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.070924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173190 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173269 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hd5r\" (UniqueName: \"kubernetes.io/projected/da76f156-fbfc-46fe-b11d-a71a0bfd280d-kube-api-access-7hd5r\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173393 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173445 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.173692 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.174498 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.178115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.178258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.179389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.179471 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.180551 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.181919 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.182194 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.191078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hd5r\" (UniqueName: \"kubernetes.io/projected/da76f156-fbfc-46fe-b11d-a71a0bfd280d-kube-api-access-7hd5r\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8rpw6\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.329635 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 08:57:15 crc kubenswrapper[4744]: I1201 08:57:15.872706 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6"] Dec 01 08:57:16 crc kubenswrapper[4744]: I1201 08:57:16.888961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" event={"ID":"da76f156-fbfc-46fe-b11d-a71a0bfd280d","Type":"ContainerStarted","Data":"bccb68356f3f04d4f127eb0d2a323c9233fcf7e457672cc984db61227c138280"} Dec 01 08:57:16 crc kubenswrapper[4744]: I1201 08:57:16.889316 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" event={"ID":"da76f156-fbfc-46fe-b11d-a71a0bfd280d","Type":"ContainerStarted","Data":"01e9cfdbbfd382fcaf879ed11aa9f743a45ac42c1abe2b723d175dd38b487ee8"} Dec 01 08:57:16 crc kubenswrapper[4744]: I1201 08:57:16.913922 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" podStartSLOduration=2.358708131 podStartE2EDuration="2.913894923s" podCreationTimestamp="2025-12-01 08:57:14 +0000 UTC" firstStartedPulling="2025-12-01 08:57:15.877723979 +0000 UTC m=+2367.866781900" lastFinishedPulling="2025-12-01 08:57:16.432910741 +0000 UTC m=+2368.421968692" observedRunningTime="2025-12-01 08:57:16.904831732 +0000 UTC m=+2368.893889663" watchObservedRunningTime="2025-12-01 08:57:16.913894923 +0000 UTC m=+2368.902952844" Dec 01 08:57:17 crc kubenswrapper[4744]: I1201 08:57:17.285679 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:57:17 crc kubenswrapper[4744]: E1201 08:57:17.285966 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:57:30 crc kubenswrapper[4744]: I1201 08:57:30.286677 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:57:30 crc kubenswrapper[4744]: E1201 08:57:30.288203 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:57:41 crc kubenswrapper[4744]: I1201 08:57:41.286898 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:57:41 crc kubenswrapper[4744]: E1201 08:57:41.313483 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:57:55 crc kubenswrapper[4744]: I1201 08:57:55.285187 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:57:55 crc kubenswrapper[4744]: E1201 08:57:55.286048 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:58:10 crc kubenswrapper[4744]: I1201 08:58:10.285179 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:58:10 crc kubenswrapper[4744]: E1201 08:58:10.288127 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:58:25 crc kubenswrapper[4744]: I1201 08:58:25.285561 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:58:25 crc kubenswrapper[4744]: E1201 08:58:25.287336 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:58:36 crc kubenswrapper[4744]: I1201 08:58:36.286338 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:58:36 crc kubenswrapper[4744]: E1201 08:58:36.286936 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:58:48 crc kubenswrapper[4744]: I1201 08:58:48.295643 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:58:48 crc kubenswrapper[4744]: E1201 08:58:48.296786 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:58:59 crc kubenswrapper[4744]: I1201 08:58:59.285702 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:58:59 crc kubenswrapper[4744]: E1201 08:58:59.286717 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:59:10 crc kubenswrapper[4744]: I1201 08:59:10.285323 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:59:10 crc kubenswrapper[4744]: E1201 08:59:10.286306 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:59:23 crc kubenswrapper[4744]: I1201 08:59:23.285039 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:59:23 crc kubenswrapper[4744]: E1201 08:59:23.285976 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:59:38 crc kubenswrapper[4744]: I1201 08:59:38.299028 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:59:38 crc kubenswrapper[4744]: E1201 08:59:38.299957 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:59:50 crc kubenswrapper[4744]: I1201 08:59:50.286336 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 08:59:50 crc kubenswrapper[4744]: E1201 08:59:50.288914 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.106246 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tjhwv"] Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.110362 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.138032 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tjhwv"] Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.257442 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-catalog-content\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.257685 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z9cc\" (UniqueName: \"kubernetes.io/projected/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-kube-api-access-8z9cc\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.257777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-utilities\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.359553 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-catalog-content\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.359746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z9cc\" (UniqueName: \"kubernetes.io/projected/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-kube-api-access-8z9cc\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.359779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-utilities\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.359956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-catalog-content\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.360322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-utilities\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.388831 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z9cc\" (UniqueName: \"kubernetes.io/projected/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-kube-api-access-8z9cc\") pod \"certified-operators-tjhwv\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.439508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 08:59:55 crc kubenswrapper[4744]: I1201 08:59:55.953240 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tjhwv"] Dec 01 08:59:56 crc kubenswrapper[4744]: I1201 08:59:56.023888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjhwv" event={"ID":"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb","Type":"ContainerStarted","Data":"159f6e39329c35628a321ad2264046081d242b264dc58bf8b46935b40d081d90"} Dec 01 08:59:57 crc kubenswrapper[4744]: I1201 08:59:57.034468 4744 generic.go:334] "Generic (PLEG): container finished" podID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerID="d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d" exitCode=0 Dec 01 08:59:57 crc kubenswrapper[4744]: I1201 08:59:57.034554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjhwv" event={"ID":"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb","Type":"ContainerDied","Data":"d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d"} Dec 01 08:59:58 crc kubenswrapper[4744]: I1201 08:59:58.046243 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjhwv" event={"ID":"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb","Type":"ContainerStarted","Data":"31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098"} Dec 01 08:59:59 crc kubenswrapper[4744]: I1201 08:59:59.062571 4744 generic.go:334] "Generic (PLEG): container finished" podID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerID="31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098" exitCode=0 Dec 01 08:59:59 crc kubenswrapper[4744]: I1201 08:59:59.062615 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjhwv" event={"ID":"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb","Type":"ContainerDied","Data":"31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098"} Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.077491 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjhwv" event={"ID":"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb","Type":"ContainerStarted","Data":"415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038"} Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.095861 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tjhwv" podStartSLOduration=2.636156374 podStartE2EDuration="5.095842783s" podCreationTimestamp="2025-12-01 08:59:55 +0000 UTC" firstStartedPulling="2025-12-01 08:59:57.038449428 +0000 UTC m=+2529.027507389" lastFinishedPulling="2025-12-01 08:59:59.498135847 +0000 UTC m=+2531.487193798" observedRunningTime="2025-12-01 09:00:00.093541417 +0000 UTC m=+2532.082599358" watchObservedRunningTime="2025-12-01 09:00:00.095842783 +0000 UTC m=+2532.084900704" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.146162 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g"] Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.147647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.149749 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.149954 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.155794 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g"] Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.265301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67ww7\" (UniqueName: \"kubernetes.io/projected/a0b4271f-9683-4cf3-a974-52e4906a5d13-kube-api-access-67ww7\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.265373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0b4271f-9683-4cf3-a974-52e4906a5d13-config-volume\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.265402 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0b4271f-9683-4cf3-a974-52e4906a5d13-secret-volume\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.368025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67ww7\" (UniqueName: \"kubernetes.io/projected/a0b4271f-9683-4cf3-a974-52e4906a5d13-kube-api-access-67ww7\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.368095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0b4271f-9683-4cf3-a974-52e4906a5d13-config-volume\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.368115 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0b4271f-9683-4cf3-a974-52e4906a5d13-secret-volume\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.369395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0b4271f-9683-4cf3-a974-52e4906a5d13-config-volume\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.375260 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0b4271f-9683-4cf3-a974-52e4906a5d13-secret-volume\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.385016 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67ww7\" (UniqueName: \"kubernetes.io/projected/a0b4271f-9683-4cf3-a974-52e4906a5d13-kube-api-access-67ww7\") pod \"collect-profiles-29409660-qrv8g\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.479150 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:00 crc kubenswrapper[4744]: I1201 09:00:00.943022 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g"] Dec 01 09:00:01 crc kubenswrapper[4744]: I1201 09:00:01.090559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" event={"ID":"a0b4271f-9683-4cf3-a974-52e4906a5d13","Type":"ContainerStarted","Data":"8b389470882402659d880d800df54d8cf17a431cff20c598c802b44e925f2f2b"} Dec 01 09:00:01 crc kubenswrapper[4744]: I1201 09:00:01.284879 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:00:01 crc kubenswrapper[4744]: E1201 09:00:01.285406 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:00:02 crc kubenswrapper[4744]: I1201 09:00:02.105466 4744 generic.go:334] "Generic (PLEG): container finished" podID="a0b4271f-9683-4cf3-a974-52e4906a5d13" containerID="d9a57554e900cd7fbc51c1838c7a807b109056f60bb04505f1fea438570f1dc2" exitCode=0 Dec 01 09:00:02 crc kubenswrapper[4744]: I1201 09:00:02.105534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" event={"ID":"a0b4271f-9683-4cf3-a974-52e4906a5d13","Type":"ContainerDied","Data":"d9a57554e900cd7fbc51c1838c7a807b109056f60bb04505f1fea438570f1dc2"} Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.492452 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.630066 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0b4271f-9683-4cf3-a974-52e4906a5d13-secret-volume\") pod \"a0b4271f-9683-4cf3-a974-52e4906a5d13\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.630153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0b4271f-9683-4cf3-a974-52e4906a5d13-config-volume\") pod \"a0b4271f-9683-4cf3-a974-52e4906a5d13\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.630305 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67ww7\" (UniqueName: \"kubernetes.io/projected/a0b4271f-9683-4cf3-a974-52e4906a5d13-kube-api-access-67ww7\") pod \"a0b4271f-9683-4cf3-a974-52e4906a5d13\" (UID: \"a0b4271f-9683-4cf3-a974-52e4906a5d13\") " Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.631239 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0b4271f-9683-4cf3-a974-52e4906a5d13-config-volume" (OuterVolumeSpecName: "config-volume") pod "a0b4271f-9683-4cf3-a974-52e4906a5d13" (UID: "a0b4271f-9683-4cf3-a974-52e4906a5d13"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.635996 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b4271f-9683-4cf3-a974-52e4906a5d13-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a0b4271f-9683-4cf3-a974-52e4906a5d13" (UID: "a0b4271f-9683-4cf3-a974-52e4906a5d13"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.636845 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b4271f-9683-4cf3-a974-52e4906a5d13-kube-api-access-67ww7" (OuterVolumeSpecName: "kube-api-access-67ww7") pod "a0b4271f-9683-4cf3-a974-52e4906a5d13" (UID: "a0b4271f-9683-4cf3-a974-52e4906a5d13"). InnerVolumeSpecName "kube-api-access-67ww7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.732842 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67ww7\" (UniqueName: \"kubernetes.io/projected/a0b4271f-9683-4cf3-a974-52e4906a5d13-kube-api-access-67ww7\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.732901 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0b4271f-9683-4cf3-a974-52e4906a5d13-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:03 crc kubenswrapper[4744]: I1201 09:00:03.732914 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0b4271f-9683-4cf3-a974-52e4906a5d13-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:04 crc kubenswrapper[4744]: I1201 09:00:04.126461 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" event={"ID":"a0b4271f-9683-4cf3-a974-52e4906a5d13","Type":"ContainerDied","Data":"8b389470882402659d880d800df54d8cf17a431cff20c598c802b44e925f2f2b"} Dec 01 09:00:04 crc kubenswrapper[4744]: I1201 09:00:04.126504 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b389470882402659d880d800df54d8cf17a431cff20c598c802b44e925f2f2b" Dec 01 09:00:04 crc kubenswrapper[4744]: I1201 09:00:04.126516 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-qrv8g" Dec 01 09:00:04 crc kubenswrapper[4744]: I1201 09:00:04.573466 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55"] Dec 01 09:00:04 crc kubenswrapper[4744]: I1201 09:00:04.582842 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409615-l7t55"] Dec 01 09:00:05 crc kubenswrapper[4744]: I1201 09:00:05.441375 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 09:00:05 crc kubenswrapper[4744]: I1201 09:00:05.441485 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 09:00:05 crc kubenswrapper[4744]: I1201 09:00:05.528255 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 09:00:06 crc kubenswrapper[4744]: I1201 09:00:06.208335 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 09:00:06 crc kubenswrapper[4744]: I1201 09:00:06.270238 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tjhwv"] Dec 01 09:00:06 crc kubenswrapper[4744]: I1201 09:00:06.307056 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3" path="/var/lib/kubelet/pods/9a34de54-62a8-4fe8-b5a6-f6541cbb4fd3/volumes" Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.172473 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tjhwv" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="registry-server" containerID="cri-o://415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038" gracePeriod=2 Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.704794 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.751733 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-catalog-content\") pod \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.752037 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-utilities\") pod \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.752073 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z9cc\" (UniqueName: \"kubernetes.io/projected/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-kube-api-access-8z9cc\") pod \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\" (UID: \"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb\") " Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.756687 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-utilities" (OuterVolumeSpecName: "utilities") pod "f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" (UID: "f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.759847 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-kube-api-access-8z9cc" (OuterVolumeSpecName: "kube-api-access-8z9cc") pod "f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" (UID: "f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb"). InnerVolumeSpecName "kube-api-access-8z9cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.802026 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" (UID: "f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.854696 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.854788 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z9cc\" (UniqueName: \"kubernetes.io/projected/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-kube-api-access-8z9cc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:08 crc kubenswrapper[4744]: I1201 09:00:08.854800 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.185634 4744 generic.go:334] "Generic (PLEG): container finished" podID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerID="415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038" exitCode=0 Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.185709 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjhwv" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.185716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjhwv" event={"ID":"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb","Type":"ContainerDied","Data":"415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038"} Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.185777 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjhwv" event={"ID":"f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb","Type":"ContainerDied","Data":"159f6e39329c35628a321ad2264046081d242b264dc58bf8b46935b40d081d90"} Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.185804 4744 scope.go:117] "RemoveContainer" containerID="415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.206422 4744 scope.go:117] "RemoveContainer" containerID="31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.221475 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tjhwv"] Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.230530 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tjhwv"] Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.239467 4744 scope.go:117] "RemoveContainer" containerID="d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.266952 4744 scope.go:117] "RemoveContainer" containerID="415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038" Dec 01 09:00:09 crc kubenswrapper[4744]: E1201 09:00:09.267320 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038\": container with ID starting with 415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038 not found: ID does not exist" containerID="415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.267355 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038"} err="failed to get container status \"415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038\": rpc error: code = NotFound desc = could not find container \"415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038\": container with ID starting with 415728a5b03c15974b4415ab7ff285eba21c38f8e36e7fee73c125aedbbe7038 not found: ID does not exist" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.267380 4744 scope.go:117] "RemoveContainer" containerID="31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098" Dec 01 09:00:09 crc kubenswrapper[4744]: E1201 09:00:09.267650 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098\": container with ID starting with 31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098 not found: ID does not exist" containerID="31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.267677 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098"} err="failed to get container status \"31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098\": rpc error: code = NotFound desc = could not find container \"31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098\": container with ID starting with 31da3d1b180f1f42612e4d0ff6740a4c755b87e6ffbc9cd810ac91e2ae4b5098 not found: ID does not exist" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.267693 4744 scope.go:117] "RemoveContainer" containerID="d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d" Dec 01 09:00:09 crc kubenswrapper[4744]: E1201 09:00:09.267939 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d\": container with ID starting with d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d not found: ID does not exist" containerID="d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d" Dec 01 09:00:09 crc kubenswrapper[4744]: I1201 09:00:09.267966 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d"} err="failed to get container status \"d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d\": rpc error: code = NotFound desc = could not find container \"d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d\": container with ID starting with d463eb17804f0ab9a309edb3777b0c5c96f2f8c4e4ff2fcdbf906cbc54ad331d not found: ID does not exist" Dec 01 09:00:10 crc kubenswrapper[4744]: I1201 09:00:10.302372 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" path="/var/lib/kubelet/pods/f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb/volumes" Dec 01 09:00:14 crc kubenswrapper[4744]: I1201 09:00:14.285798 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:00:14 crc kubenswrapper[4744]: E1201 09:00:14.286761 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:00:24 crc kubenswrapper[4744]: I1201 09:00:24.326270 4744 generic.go:334] "Generic (PLEG): container finished" podID="da76f156-fbfc-46fe-b11d-a71a0bfd280d" containerID="bccb68356f3f04d4f127eb0d2a323c9233fcf7e457672cc984db61227c138280" exitCode=0 Dec 01 09:00:24 crc kubenswrapper[4744]: I1201 09:00:24.326346 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" event={"ID":"da76f156-fbfc-46fe-b11d-a71a0bfd280d","Type":"ContainerDied","Data":"bccb68356f3f04d4f127eb0d2a323c9233fcf7e457672cc984db61227c138280"} Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.762167 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886193 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-0\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-extra-config-0\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886288 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-0\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886311 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-1\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886349 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-1\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886367 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-ssh-key\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886421 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hd5r\" (UniqueName: \"kubernetes.io/projected/da76f156-fbfc-46fe-b11d-a71a0bfd280d-kube-api-access-7hd5r\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886498 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-inventory\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.886516 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-combined-ca-bundle\") pod \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\" (UID: \"da76f156-fbfc-46fe-b11d-a71a0bfd280d\") " Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.892845 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da76f156-fbfc-46fe-b11d-a71a0bfd280d-kube-api-access-7hd5r" (OuterVolumeSpecName: "kube-api-access-7hd5r") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "kube-api-access-7hd5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.892917 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.915493 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.916978 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.919418 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-inventory" (OuterVolumeSpecName: "inventory") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.921236 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.923514 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.924514 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.927686 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "da76f156-fbfc-46fe-b11d-a71a0bfd280d" (UID: "da76f156-fbfc-46fe-b11d-a71a0bfd280d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989219 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989690 4744 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989702 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989711 4744 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989760 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989771 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989779 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989787 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da76f156-fbfc-46fe-b11d-a71a0bfd280d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:25 crc kubenswrapper[4744]: I1201 09:00:25.989796 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hd5r\" (UniqueName: \"kubernetes.io/projected/da76f156-fbfc-46fe-b11d-a71a0bfd280d-kube-api-access-7hd5r\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.285598 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:00:26 crc kubenswrapper[4744]: E1201 09:00:26.285919 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.387690 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.390123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8rpw6" event={"ID":"da76f156-fbfc-46fe-b11d-a71a0bfd280d","Type":"ContainerDied","Data":"01e9cfdbbfd382fcaf879ed11aa9f743a45ac42c1abe2b723d175dd38b487ee8"} Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.390185 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01e9cfdbbfd382fcaf879ed11aa9f743a45ac42c1abe2b723d175dd38b487ee8" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.466590 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv"] Dec 01 09:00:26 crc kubenswrapper[4744]: E1201 09:00:26.467040 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="extract-utilities" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467061 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="extract-utilities" Dec 01 09:00:26 crc kubenswrapper[4744]: E1201 09:00:26.467081 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="registry-server" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467089 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="registry-server" Dec 01 09:00:26 crc kubenswrapper[4744]: E1201 09:00:26.467108 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da76f156-fbfc-46fe-b11d-a71a0bfd280d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467116 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="da76f156-fbfc-46fe-b11d-a71a0bfd280d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 09:00:26 crc kubenswrapper[4744]: E1201 09:00:26.467144 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b4271f-9683-4cf3-a974-52e4906a5d13" containerName="collect-profiles" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467152 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b4271f-9683-4cf3-a974-52e4906a5d13" containerName="collect-profiles" Dec 01 09:00:26 crc kubenswrapper[4744]: E1201 09:00:26.467191 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="extract-content" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467199 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="extract-content" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467442 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f60b0a52-35ff-4762-aa7a-ffe3e0f96cbb" containerName="registry-server" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467465 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b4271f-9683-4cf3-a974-52e4906a5d13" containerName="collect-profiles" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.467480 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="da76f156-fbfc-46fe-b11d-a71a0bfd280d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.468183 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.470271 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6bn9m" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.470875 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.471131 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.471315 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.475486 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.476089 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv"] Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.604290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.604499 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.604564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.604613 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.604640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.604682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wff7d\" (UniqueName: \"kubernetes.io/projected/ffbaa140-8ee6-4639-9524-96460ffe5a49-kube-api-access-wff7d\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.604759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.706537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.706637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.706670 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.706703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wff7d\" (UniqueName: \"kubernetes.io/projected/ffbaa140-8ee6-4639-9524-96460ffe5a49-kube-api-access-wff7d\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.706774 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.706827 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.706916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.710905 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.711047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.711508 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.711671 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.713015 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.721203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.723039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wff7d\" (UniqueName: \"kubernetes.io/projected/ffbaa140-8ee6-4639-9524-96460ffe5a49-kube-api-access-wff7d\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s24rv\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:26 crc kubenswrapper[4744]: I1201 09:00:26.801436 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:00:27 crc kubenswrapper[4744]: I1201 09:00:27.332721 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv"] Dec 01 09:00:27 crc kubenswrapper[4744]: I1201 09:00:27.335280 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:00:27 crc kubenswrapper[4744]: I1201 09:00:27.396271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" event={"ID":"ffbaa140-8ee6-4639-9524-96460ffe5a49","Type":"ContainerStarted","Data":"79d8a0a01f58fe4e175f221333037782188534f6843079bd1cfa149c7e4cbb8e"} Dec 01 09:00:28 crc kubenswrapper[4744]: I1201 09:00:28.411798 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" event={"ID":"ffbaa140-8ee6-4639-9524-96460ffe5a49","Type":"ContainerStarted","Data":"c40925ef957b2b15621ff99a7853e1930b6ceea599bad1fed2467e6d601a8702"} Dec 01 09:00:28 crc kubenswrapper[4744]: I1201 09:00:28.440754 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" podStartSLOduration=1.82500208 podStartE2EDuration="2.440733843s" podCreationTimestamp="2025-12-01 09:00:26 +0000 UTC" firstStartedPulling="2025-12-01 09:00:27.335070574 +0000 UTC m=+2559.324128495" lastFinishedPulling="2025-12-01 09:00:27.950802317 +0000 UTC m=+2559.939860258" observedRunningTime="2025-12-01 09:00:28.437741728 +0000 UTC m=+2560.426799669" watchObservedRunningTime="2025-12-01 09:00:28.440733843 +0000 UTC m=+2560.429791764" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.060099 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l8j9f"] Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.063133 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.078252 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l8j9f"] Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.207072 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-utilities\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.207121 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbk6w\" (UniqueName: \"kubernetes.io/projected/b38c6827-3551-428c-b4a9-a4600177ebd8-kube-api-access-hbk6w\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.207152 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-catalog-content\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.308639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-utilities\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.308692 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbk6w\" (UniqueName: \"kubernetes.io/projected/b38c6827-3551-428c-b4a9-a4600177ebd8-kube-api-access-hbk6w\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.309566 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-utilities\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.310000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-catalog-content\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.310350 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-catalog-content\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.341354 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbk6w\" (UniqueName: \"kubernetes.io/projected/b38c6827-3551-428c-b4a9-a4600177ebd8-kube-api-access-hbk6w\") pod \"community-operators-l8j9f\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.383920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:36 crc kubenswrapper[4744]: I1201 09:00:36.897505 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l8j9f"] Dec 01 09:00:36 crc kubenswrapper[4744]: W1201 09:00:36.902381 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb38c6827_3551_428c_b4a9_a4600177ebd8.slice/crio-c6375861f890fd164cb9870835fe1d832dd32f341089fd9f11fda243c15e8cfc WatchSource:0}: Error finding container c6375861f890fd164cb9870835fe1d832dd32f341089fd9f11fda243c15e8cfc: Status 404 returned error can't find the container with id c6375861f890fd164cb9870835fe1d832dd32f341089fd9f11fda243c15e8cfc Dec 01 09:00:37 crc kubenswrapper[4744]: I1201 09:00:37.507154 4744 generic.go:334] "Generic (PLEG): container finished" podID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerID="94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85" exitCode=0 Dec 01 09:00:37 crc kubenswrapper[4744]: I1201 09:00:37.507240 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8j9f" event={"ID":"b38c6827-3551-428c-b4a9-a4600177ebd8","Type":"ContainerDied","Data":"94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85"} Dec 01 09:00:37 crc kubenswrapper[4744]: I1201 09:00:37.507417 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8j9f" event={"ID":"b38c6827-3551-428c-b4a9-a4600177ebd8","Type":"ContainerStarted","Data":"c6375861f890fd164cb9870835fe1d832dd32f341089fd9f11fda243c15e8cfc"} Dec 01 09:00:39 crc kubenswrapper[4744]: I1201 09:00:39.533194 4744 generic.go:334] "Generic (PLEG): container finished" podID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerID="85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058" exitCode=0 Dec 01 09:00:39 crc kubenswrapper[4744]: I1201 09:00:39.533602 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8j9f" event={"ID":"b38c6827-3551-428c-b4a9-a4600177ebd8","Type":"ContainerDied","Data":"85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058"} Dec 01 09:00:40 crc kubenswrapper[4744]: I1201 09:00:40.284898 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:00:40 crc kubenswrapper[4744]: E1201 09:00:40.285515 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:00:40 crc kubenswrapper[4744]: I1201 09:00:40.545003 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8j9f" event={"ID":"b38c6827-3551-428c-b4a9-a4600177ebd8","Type":"ContainerStarted","Data":"e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057"} Dec 01 09:00:40 crc kubenswrapper[4744]: I1201 09:00:40.561291 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l8j9f" podStartSLOduration=1.934268333 podStartE2EDuration="4.56127149s" podCreationTimestamp="2025-12-01 09:00:36 +0000 UTC" firstStartedPulling="2025-12-01 09:00:37.508912068 +0000 UTC m=+2569.497969989" lastFinishedPulling="2025-12-01 09:00:40.135915225 +0000 UTC m=+2572.124973146" observedRunningTime="2025-12-01 09:00:40.558695466 +0000 UTC m=+2572.547753397" watchObservedRunningTime="2025-12-01 09:00:40.56127149 +0000 UTC m=+2572.550329411" Dec 01 09:00:46 crc kubenswrapper[4744]: I1201 09:00:46.384810 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:46 crc kubenswrapper[4744]: I1201 09:00:46.385104 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:46 crc kubenswrapper[4744]: I1201 09:00:46.448123 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:46 crc kubenswrapper[4744]: I1201 09:00:46.673064 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:46 crc kubenswrapper[4744]: I1201 09:00:46.712940 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l8j9f"] Dec 01 09:00:48 crc kubenswrapper[4744]: I1201 09:00:48.636608 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l8j9f" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="registry-server" containerID="cri-o://e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057" gracePeriod=2 Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.118662 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.272348 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-catalog-content\") pod \"b38c6827-3551-428c-b4a9-a4600177ebd8\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.272887 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbk6w\" (UniqueName: \"kubernetes.io/projected/b38c6827-3551-428c-b4a9-a4600177ebd8-kube-api-access-hbk6w\") pod \"b38c6827-3551-428c-b4a9-a4600177ebd8\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.272930 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-utilities\") pod \"b38c6827-3551-428c-b4a9-a4600177ebd8\" (UID: \"b38c6827-3551-428c-b4a9-a4600177ebd8\") " Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.273939 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-utilities" (OuterVolumeSpecName: "utilities") pod "b38c6827-3551-428c-b4a9-a4600177ebd8" (UID: "b38c6827-3551-428c-b4a9-a4600177ebd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.284582 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b38c6827-3551-428c-b4a9-a4600177ebd8-kube-api-access-hbk6w" (OuterVolumeSpecName: "kube-api-access-hbk6w") pod "b38c6827-3551-428c-b4a9-a4600177ebd8" (UID: "b38c6827-3551-428c-b4a9-a4600177ebd8"). InnerVolumeSpecName "kube-api-access-hbk6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.324379 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b38c6827-3551-428c-b4a9-a4600177ebd8" (UID: "b38c6827-3551-428c-b4a9-a4600177ebd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.375892 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.375925 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b38c6827-3551-428c-b4a9-a4600177ebd8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.375937 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbk6w\" (UniqueName: \"kubernetes.io/projected/b38c6827-3551-428c-b4a9-a4600177ebd8-kube-api-access-hbk6w\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.646907 4744 generic.go:334] "Generic (PLEG): container finished" podID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerID="e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057" exitCode=0 Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.646950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8j9f" event={"ID":"b38c6827-3551-428c-b4a9-a4600177ebd8","Type":"ContainerDied","Data":"e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057"} Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.646975 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l8j9f" event={"ID":"b38c6827-3551-428c-b4a9-a4600177ebd8","Type":"ContainerDied","Data":"c6375861f890fd164cb9870835fe1d832dd32f341089fd9f11fda243c15e8cfc"} Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.646991 4744 scope.go:117] "RemoveContainer" containerID="e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.647113 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l8j9f" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.676445 4744 scope.go:117] "RemoveContainer" containerID="85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.678901 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l8j9f"] Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.689004 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l8j9f"] Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.699204 4744 scope.go:117] "RemoveContainer" containerID="94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.761523 4744 scope.go:117] "RemoveContainer" containerID="e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057" Dec 01 09:00:49 crc kubenswrapper[4744]: E1201 09:00:49.761927 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057\": container with ID starting with e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057 not found: ID does not exist" containerID="e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.761987 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057"} err="failed to get container status \"e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057\": rpc error: code = NotFound desc = could not find container \"e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057\": container with ID starting with e3b2cca7847036a154519d61dae32b73e2726009c7155ce91e5581e645294057 not found: ID does not exist" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.762027 4744 scope.go:117] "RemoveContainer" containerID="85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058" Dec 01 09:00:49 crc kubenswrapper[4744]: E1201 09:00:49.762337 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058\": container with ID starting with 85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058 not found: ID does not exist" containerID="85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.762374 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058"} err="failed to get container status \"85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058\": rpc error: code = NotFound desc = could not find container \"85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058\": container with ID starting with 85fd36857962b86d510ffb89a73facc1029dc42ffca75d50699845cb72084058 not found: ID does not exist" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.762400 4744 scope.go:117] "RemoveContainer" containerID="94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85" Dec 01 09:00:49 crc kubenswrapper[4744]: E1201 09:00:49.762708 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85\": container with ID starting with 94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85 not found: ID does not exist" containerID="94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85" Dec 01 09:00:49 crc kubenswrapper[4744]: I1201 09:00:49.762753 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85"} err="failed to get container status \"94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85\": rpc error: code = NotFound desc = could not find container \"94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85\": container with ID starting with 94cbc1c28c63a2dcefc08071bd872f6f92191d5910e7e67341519db9788ace85 not found: ID does not exist" Dec 01 09:00:50 crc kubenswrapper[4744]: I1201 09:00:50.297393 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" path="/var/lib/kubelet/pods/b38c6827-3551-428c-b4a9-a4600177ebd8/volumes" Dec 01 09:00:51 crc kubenswrapper[4744]: I1201 09:00:51.285256 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:00:51 crc kubenswrapper[4744]: E1201 09:00:51.285848 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.156993 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29409661-lht94"] Dec 01 09:01:00 crc kubenswrapper[4744]: E1201 09:01:00.158067 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="registry-server" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.158086 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="registry-server" Dec 01 09:01:00 crc kubenswrapper[4744]: E1201 09:01:00.158122 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="extract-content" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.158132 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="extract-content" Dec 01 09:01:00 crc kubenswrapper[4744]: E1201 09:01:00.158151 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="extract-utilities" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.158160 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="extract-utilities" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.158456 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38c6827-3551-428c-b4a9-a4600177ebd8" containerName="registry-server" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.159322 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.167372 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29409661-lht94"] Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.204559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnfv6\" (UniqueName: \"kubernetes.io/projected/ee0b29aa-22fe-412e-ad54-43db39412cec-kube-api-access-nnfv6\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.204825 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-fernet-keys\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.204908 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-combined-ca-bundle\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.204971 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-config-data\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.307272 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnfv6\" (UniqueName: \"kubernetes.io/projected/ee0b29aa-22fe-412e-ad54-43db39412cec-kube-api-access-nnfv6\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.307523 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-fernet-keys\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.307570 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-combined-ca-bundle\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.307611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-config-data\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.317190 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-combined-ca-bundle\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.317306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-fernet-keys\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.317697 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-config-data\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.325191 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnfv6\" (UniqueName: \"kubernetes.io/projected/ee0b29aa-22fe-412e-ad54-43db39412cec-kube-api-access-nnfv6\") pod \"keystone-cron-29409661-lht94\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.488717 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:00 crc kubenswrapper[4744]: I1201 09:01:00.911755 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29409661-lht94"] Dec 01 09:01:00 crc kubenswrapper[4744]: W1201 09:01:00.919055 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee0b29aa_22fe_412e_ad54_43db39412cec.slice/crio-27ee1712503691652b8086695050813ecb4debfebf505f9bf4365c6f87e37db7 WatchSource:0}: Error finding container 27ee1712503691652b8086695050813ecb4debfebf505f9bf4365c6f87e37db7: Status 404 returned error can't find the container with id 27ee1712503691652b8086695050813ecb4debfebf505f9bf4365c6f87e37db7 Dec 01 09:01:01 crc kubenswrapper[4744]: I1201 09:01:01.774801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-lht94" event={"ID":"ee0b29aa-22fe-412e-ad54-43db39412cec","Type":"ContainerStarted","Data":"26592a37b6316376db034bdac4a716702c6a5dedee850519ae248242bc445086"} Dec 01 09:01:01 crc kubenswrapper[4744]: I1201 09:01:01.775135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-lht94" event={"ID":"ee0b29aa-22fe-412e-ad54-43db39412cec","Type":"ContainerStarted","Data":"27ee1712503691652b8086695050813ecb4debfebf505f9bf4365c6f87e37db7"} Dec 01 09:01:01 crc kubenswrapper[4744]: I1201 09:01:01.796747 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29409661-lht94" podStartSLOduration=1.796729053 podStartE2EDuration="1.796729053s" podCreationTimestamp="2025-12-01 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:01.790295268 +0000 UTC m=+2593.779353199" watchObservedRunningTime="2025-12-01 09:01:01.796729053 +0000 UTC m=+2593.785786984" Dec 01 09:01:03 crc kubenswrapper[4744]: I1201 09:01:03.726940 4744 scope.go:117] "RemoveContainer" containerID="1c84030850d7c387ba2b783090dbdd76b32f8ba6415f8fb7eb24e007f6a6e33a" Dec 01 09:01:03 crc kubenswrapper[4744]: I1201 09:01:03.796055 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee0b29aa-22fe-412e-ad54-43db39412cec" containerID="26592a37b6316376db034bdac4a716702c6a5dedee850519ae248242bc445086" exitCode=0 Dec 01 09:01:03 crc kubenswrapper[4744]: I1201 09:01:03.796118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-lht94" event={"ID":"ee0b29aa-22fe-412e-ad54-43db39412cec","Type":"ContainerDied","Data":"26592a37b6316376db034bdac4a716702c6a5dedee850519ae248242bc445086"} Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.109474 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.285185 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:01:05 crc kubenswrapper[4744]: E1201 09:01:05.285514 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.302101 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-fernet-keys\") pod \"ee0b29aa-22fe-412e-ad54-43db39412cec\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.302247 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-combined-ca-bundle\") pod \"ee0b29aa-22fe-412e-ad54-43db39412cec\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.302736 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-config-data\") pod \"ee0b29aa-22fe-412e-ad54-43db39412cec\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.302799 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnfv6\" (UniqueName: \"kubernetes.io/projected/ee0b29aa-22fe-412e-ad54-43db39412cec-kube-api-access-nnfv6\") pod \"ee0b29aa-22fe-412e-ad54-43db39412cec\" (UID: \"ee0b29aa-22fe-412e-ad54-43db39412cec\") " Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.308717 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0b29aa-22fe-412e-ad54-43db39412cec-kube-api-access-nnfv6" (OuterVolumeSpecName: "kube-api-access-nnfv6") pod "ee0b29aa-22fe-412e-ad54-43db39412cec" (UID: "ee0b29aa-22fe-412e-ad54-43db39412cec"). InnerVolumeSpecName "kube-api-access-nnfv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.309658 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ee0b29aa-22fe-412e-ad54-43db39412cec" (UID: "ee0b29aa-22fe-412e-ad54-43db39412cec"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.334118 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee0b29aa-22fe-412e-ad54-43db39412cec" (UID: "ee0b29aa-22fe-412e-ad54-43db39412cec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.355789 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-config-data" (OuterVolumeSpecName: "config-data") pod "ee0b29aa-22fe-412e-ad54-43db39412cec" (UID: "ee0b29aa-22fe-412e-ad54-43db39412cec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.405622 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.405673 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.405683 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnfv6\" (UniqueName: \"kubernetes.io/projected/ee0b29aa-22fe-412e-ad54-43db39412cec-kube-api-access-nnfv6\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.405692 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee0b29aa-22fe-412e-ad54-43db39412cec-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.814525 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29409661-lht94" event={"ID":"ee0b29aa-22fe-412e-ad54-43db39412cec","Type":"ContainerDied","Data":"27ee1712503691652b8086695050813ecb4debfebf505f9bf4365c6f87e37db7"} Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.814567 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27ee1712503691652b8086695050813ecb4debfebf505f9bf4365c6f87e37db7" Dec 01 09:01:05 crc kubenswrapper[4744]: I1201 09:01:05.814587 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29409661-lht94" Dec 01 09:01:19 crc kubenswrapper[4744]: I1201 09:01:19.286396 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:01:19 crc kubenswrapper[4744]: I1201 09:01:19.992203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"23b9d98ac18b222c3a666b71f499936c00e4bbbeaa0860bc73299c7c08de693f"} Dec 01 09:03:02 crc kubenswrapper[4744]: I1201 09:03:02.135297 4744 generic.go:334] "Generic (PLEG): container finished" podID="ffbaa140-8ee6-4639-9524-96460ffe5a49" containerID="c40925ef957b2b15621ff99a7853e1930b6ceea599bad1fed2467e6d601a8702" exitCode=0 Dec 01 09:03:02 crc kubenswrapper[4744]: I1201 09:03:02.135371 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" event={"ID":"ffbaa140-8ee6-4639-9524-96460ffe5a49","Type":"ContainerDied","Data":"c40925ef957b2b15621ff99a7853e1930b6ceea599bad1fed2467e6d601a8702"} Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.621342 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.777113 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-inventory\") pod \"ffbaa140-8ee6-4639-9524-96460ffe5a49\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.777549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-2\") pod \"ffbaa140-8ee6-4639-9524-96460ffe5a49\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.777830 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-1\") pod \"ffbaa140-8ee6-4639-9524-96460ffe5a49\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.778775 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ssh-key\") pod \"ffbaa140-8ee6-4639-9524-96460ffe5a49\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.778831 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-telemetry-combined-ca-bundle\") pod \"ffbaa140-8ee6-4639-9524-96460ffe5a49\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.778873 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-0\") pod \"ffbaa140-8ee6-4639-9524-96460ffe5a49\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.778910 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wff7d\" (UniqueName: \"kubernetes.io/projected/ffbaa140-8ee6-4639-9524-96460ffe5a49-kube-api-access-wff7d\") pod \"ffbaa140-8ee6-4639-9524-96460ffe5a49\" (UID: \"ffbaa140-8ee6-4639-9524-96460ffe5a49\") " Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.790896 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffbaa140-8ee6-4639-9524-96460ffe5a49-kube-api-access-wff7d" (OuterVolumeSpecName: "kube-api-access-wff7d") pod "ffbaa140-8ee6-4639-9524-96460ffe5a49" (UID: "ffbaa140-8ee6-4639-9524-96460ffe5a49"). InnerVolumeSpecName "kube-api-access-wff7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.790901 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ffbaa140-8ee6-4639-9524-96460ffe5a49" (UID: "ffbaa140-8ee6-4639-9524-96460ffe5a49"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.811103 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ffbaa140-8ee6-4639-9524-96460ffe5a49" (UID: "ffbaa140-8ee6-4639-9524-96460ffe5a49"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.813220 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "ffbaa140-8ee6-4639-9524-96460ffe5a49" (UID: "ffbaa140-8ee6-4639-9524-96460ffe5a49"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.829635 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "ffbaa140-8ee6-4639-9524-96460ffe5a49" (UID: "ffbaa140-8ee6-4639-9524-96460ffe5a49"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.833630 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-inventory" (OuterVolumeSpecName: "inventory") pod "ffbaa140-8ee6-4639-9524-96460ffe5a49" (UID: "ffbaa140-8ee6-4639-9524-96460ffe5a49"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.862198 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "ffbaa140-8ee6-4639-9524-96460ffe5a49" (UID: "ffbaa140-8ee6-4639-9524-96460ffe5a49"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.885172 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.885208 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.885219 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.885228 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wff7d\" (UniqueName: \"kubernetes.io/projected/ffbaa140-8ee6-4639-9524-96460ffe5a49-kube-api-access-wff7d\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.885237 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.885245 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:03 crc kubenswrapper[4744]: I1201 09:03:03.885255 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ffbaa140-8ee6-4639-9524-96460ffe5a49-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:04 crc kubenswrapper[4744]: I1201 09:03:04.157539 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" event={"ID":"ffbaa140-8ee6-4639-9524-96460ffe5a49","Type":"ContainerDied","Data":"79d8a0a01f58fe4e175f221333037782188534f6843079bd1cfa149c7e4cbb8e"} Dec 01 09:03:04 crc kubenswrapper[4744]: I1201 09:03:04.157595 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79d8a0a01f58fe4e175f221333037782188534f6843079bd1cfa149c7e4cbb8e" Dec 01 09:03:04 crc kubenswrapper[4744]: I1201 09:03:04.157669 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s24rv" Dec 01 09:03:48 crc kubenswrapper[4744]: I1201 09:03:48.634534 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:03:48 crc kubenswrapper[4744]: I1201 09:03:48.636095 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:04:18 crc kubenswrapper[4744]: I1201 09:04:18.634360 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:04:18 crc kubenswrapper[4744]: I1201 09:04:18.634867 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:04:48 crc kubenswrapper[4744]: I1201 09:04:48.634048 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:04:48 crc kubenswrapper[4744]: I1201 09:04:48.634587 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:04:48 crc kubenswrapper[4744]: I1201 09:04:48.634636 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 09:04:48 crc kubenswrapper[4744]: I1201 09:04:48.635365 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"23b9d98ac18b222c3a666b71f499936c00e4bbbeaa0860bc73299c7c08de693f"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:04:48 crc kubenswrapper[4744]: I1201 09:04:48.635436 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://23b9d98ac18b222c3a666b71f499936c00e4bbbeaa0860bc73299c7c08de693f" gracePeriod=600 Dec 01 09:04:49 crc kubenswrapper[4744]: I1201 09:04:49.435018 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="23b9d98ac18b222c3a666b71f499936c00e4bbbeaa0860bc73299c7c08de693f" exitCode=0 Dec 01 09:04:49 crc kubenswrapper[4744]: I1201 09:04:49.435129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"23b9d98ac18b222c3a666b71f499936c00e4bbbeaa0860bc73299c7c08de693f"} Dec 01 09:04:49 crc kubenswrapper[4744]: I1201 09:04:49.435662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471"} Dec 01 09:04:49 crc kubenswrapper[4744]: I1201 09:04:49.435684 4744 scope.go:117] "RemoveContainer" containerID="a1657692e2533aeb444531cafcf4ad48e74bd0cfe9f230b3e635309d87cdfabc" Dec 01 09:05:58 crc kubenswrapper[4744]: I1201 09:05:58.968846 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.653234 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.654629 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="181b6931-2498-4979-a9b1-f6cbc925c78e" containerName="openstackclient" containerID="cri-o://ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b" gracePeriod=2 Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.662906 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.699452 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: E1201 09:06:00.700294 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181b6931-2498-4979-a9b1-f6cbc925c78e" containerName="openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.700318 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="181b6931-2498-4979-a9b1-f6cbc925c78e" containerName="openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: E1201 09:06:00.700346 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0b29aa-22fe-412e-ad54-43db39412cec" containerName="keystone-cron" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.700355 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0b29aa-22fe-412e-ad54-43db39412cec" containerName="keystone-cron" Dec 01 09:06:00 crc kubenswrapper[4744]: E1201 09:06:00.700397 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffbaa140-8ee6-4639-9524-96460ffe5a49" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.700425 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffbaa140-8ee6-4639-9524-96460ffe5a49" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.700853 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffbaa140-8ee6-4639-9524-96460ffe5a49" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.700904 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee0b29aa-22fe-412e-ad54-43db39412cec" containerName="keystone-cron" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.700930 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="181b6931-2498-4979-a9b1-f6cbc925c78e" containerName="openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.701749 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.731937 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.742247 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: E1201 09:06:00.743005 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-vhdzp openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="2ca48f38-1f48-4723-8048-30bd7db6a2d9" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.750460 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.757779 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.759359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.762343 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2ca48f38-1f48-4723-8048-30bd7db6a2d9" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.764155 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="181b6931-2498-4979-a9b1-f6cbc925c78e" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.766484 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.825736 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config-secret\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.826375 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.826435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.826558 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhdzp\" (UniqueName: \"kubernetes.io/projected/2ca48f38-1f48-4723-8048-30bd7db6a2d9-kube-api-access-vhdzp\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.927969 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhdzp\" (UniqueName: \"kubernetes.io/projected/2ca48f38-1f48-4723-8048-30bd7db6a2d9-kube-api-access-vhdzp\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.928093 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.928124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config-secret\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.928162 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config-secret\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.928325 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.928473 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4jwc\" (UniqueName: \"kubernetes.io/projected/ac90703c-38e1-4f7b-be3b-9b36f5467aab-kube-api-access-b4jwc\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.928575 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.928604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.929480 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: E1201 09:06:00.931188 4744 projected.go:194] Error preparing data for projected volume kube-api-access-vhdzp for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (2ca48f38-1f48-4723-8048-30bd7db6a2d9) does not match the UID in record. The object might have been deleted and then recreated Dec 01 09:06:00 crc kubenswrapper[4744]: E1201 09:06:00.931242 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2ca48f38-1f48-4723-8048-30bd7db6a2d9-kube-api-access-vhdzp podName:2ca48f38-1f48-4723-8048-30bd7db6a2d9 nodeName:}" failed. No retries permitted until 2025-12-01 09:06:01.43122362 +0000 UTC m=+2893.420281751 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vhdzp" (UniqueName: "kubernetes.io/projected/2ca48f38-1f48-4723-8048-30bd7db6a2d9-kube-api-access-vhdzp") pod "openstackclient" (UID: "2ca48f38-1f48-4723-8048-30bd7db6a2d9") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (2ca48f38-1f48-4723-8048-30bd7db6a2d9) does not match the UID in record. The object might have been deleted and then recreated Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.935093 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config-secret\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:00 crc kubenswrapper[4744]: I1201 09:06:00.935630 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.030330 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.030380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config-secret\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.030440 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.030478 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4jwc\" (UniqueName: \"kubernetes.io/projected/ac90703c-38e1-4f7b-be3b-9b36f5467aab-kube-api-access-b4jwc\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.031310 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.033793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config-secret\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.033821 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.048690 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4jwc\" (UniqueName: \"kubernetes.io/projected/ac90703c-38e1-4f7b-be3b-9b36f5467aab-kube-api-access-b4jwc\") pod \"openstackclient\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.083220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.140326 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.145900 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2ca48f38-1f48-4723-8048-30bd7db6a2d9" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.188167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.190911 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2ca48f38-1f48-4723-8048-30bd7db6a2d9" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.251015 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config-secret\") pod \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.251135 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-combined-ca-bundle\") pod \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.251204 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config\") pod \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\" (UID: \"2ca48f38-1f48-4723-8048-30bd7db6a2d9\") " Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.251774 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "2ca48f38-1f48-4723-8048-30bd7db6a2d9" (UID: "2ca48f38-1f48-4723-8048-30bd7db6a2d9"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.253761 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhdzp\" (UniqueName: \"kubernetes.io/projected/2ca48f38-1f48-4723-8048-30bd7db6a2d9-kube-api-access-vhdzp\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.253794 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.257032 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2ca48f38-1f48-4723-8048-30bd7db6a2d9" (UID: "2ca48f38-1f48-4723-8048-30bd7db6a2d9"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.260649 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ca48f38-1f48-4723-8048-30bd7db6a2d9" (UID: "2ca48f38-1f48-4723-8048-30bd7db6a2d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.356199 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.356224 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ca48f38-1f48-4723-8048-30bd7db6a2d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.638863 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.813794 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-f4krx"] Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.815314 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.825405 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-f4krx"] Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.865311 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9pck\" (UniqueName: \"kubernetes.io/projected/2d75d787-92a2-4fe3-8783-73d3029693a9-kube-api-access-t9pck\") pod \"aodh-db-create-f4krx\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.865357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d75d787-92a2-4fe3-8783-73d3029693a9-operator-scripts\") pod \"aodh-db-create-f4krx\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.921804 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-b02d-account-create-update-vkxp4"] Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.923012 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.925434 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.930626 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-b02d-account-create-update-vkxp4"] Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.969651 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c983cab-0958-4f0d-964c-1a188cad056e-operator-scripts\") pod \"aodh-b02d-account-create-update-vkxp4\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.969843 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9pck\" (UniqueName: \"kubernetes.io/projected/2d75d787-92a2-4fe3-8783-73d3029693a9-kube-api-access-t9pck\") pod \"aodh-db-create-f4krx\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.969895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d75d787-92a2-4fe3-8783-73d3029693a9-operator-scripts\") pod \"aodh-db-create-f4krx\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.969935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsbng\" (UniqueName: \"kubernetes.io/projected/6c983cab-0958-4f0d-964c-1a188cad056e-kube-api-access-nsbng\") pod \"aodh-b02d-account-create-update-vkxp4\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.970986 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d75d787-92a2-4fe3-8783-73d3029693a9-operator-scripts\") pod \"aodh-db-create-f4krx\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:01 crc kubenswrapper[4744]: I1201 09:06:01.986869 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9pck\" (UniqueName: \"kubernetes.io/projected/2d75d787-92a2-4fe3-8783-73d3029693a9-kube-api-access-t9pck\") pod \"aodh-db-create-f4krx\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.071176 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsbng\" (UniqueName: \"kubernetes.io/projected/6c983cab-0958-4f0d-964c-1a188cad056e-kube-api-access-nsbng\") pod \"aodh-b02d-account-create-update-vkxp4\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.071797 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c983cab-0958-4f0d-964c-1a188cad056e-operator-scripts\") pod \"aodh-b02d-account-create-update-vkxp4\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.073040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c983cab-0958-4f0d-964c-1a188cad056e-operator-scripts\") pod \"aodh-b02d-account-create-update-vkxp4\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.089686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsbng\" (UniqueName: \"kubernetes.io/projected/6c983cab-0958-4f0d-964c-1a188cad056e-kube-api-access-nsbng\") pod \"aodh-b02d-account-create-update-vkxp4\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.136852 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.160082 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.160109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ac90703c-38e1-4f7b-be3b-9b36f5467aab","Type":"ContainerStarted","Data":"206eec330034b3141e9e3e9d46ba68368d3d0f9a7ae83fe695fc5810d8438667"} Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.160188 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ac90703c-38e1-4f7b-be3b-9b36f5467aab","Type":"ContainerStarted","Data":"6ee6780d43328a4c0e965f6ef6bd2c671f37ac268babc2878098b7795baa4c41"} Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.190980 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2ca48f38-1f48-4723-8048-30bd7db6a2d9" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.194607 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.19458326 podStartE2EDuration="2.19458326s" podCreationTimestamp="2025-12-01 09:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:06:02.185207052 +0000 UTC m=+2894.174265003" watchObservedRunningTime="2025-12-01 09:06:02.19458326 +0000 UTC m=+2894.183641181" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.266022 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2ca48f38-1f48-4723-8048-30bd7db6a2d9" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.302530 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ca48f38-1f48-4723-8048-30bd7db6a2d9" path="/var/lib/kubelet/pods/2ca48f38-1f48-4723-8048-30bd7db6a2d9/volumes" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.348687 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.618746 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-f4krx"] Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.628364 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-b02d-account-create-update-vkxp4"] Dec 01 09:06:02 crc kubenswrapper[4744]: I1201 09:06:02.893535 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.003279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config\") pod \"181b6931-2498-4979-a9b1-f6cbc925c78e\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.003380 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmnjj\" (UniqueName: \"kubernetes.io/projected/181b6931-2498-4979-a9b1-f6cbc925c78e-kube-api-access-rmnjj\") pod \"181b6931-2498-4979-a9b1-f6cbc925c78e\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.003485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-combined-ca-bundle\") pod \"181b6931-2498-4979-a9b1-f6cbc925c78e\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.003507 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config-secret\") pod \"181b6931-2498-4979-a9b1-f6cbc925c78e\" (UID: \"181b6931-2498-4979-a9b1-f6cbc925c78e\") " Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.013344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181b6931-2498-4979-a9b1-f6cbc925c78e-kube-api-access-rmnjj" (OuterVolumeSpecName: "kube-api-access-rmnjj") pod "181b6931-2498-4979-a9b1-f6cbc925c78e" (UID: "181b6931-2498-4979-a9b1-f6cbc925c78e"). InnerVolumeSpecName "kube-api-access-rmnjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.048328 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "181b6931-2498-4979-a9b1-f6cbc925c78e" (UID: "181b6931-2498-4979-a9b1-f6cbc925c78e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.054595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "181b6931-2498-4979-a9b1-f6cbc925c78e" (UID: "181b6931-2498-4979-a9b1-f6cbc925c78e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.075008 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "181b6931-2498-4979-a9b1-f6cbc925c78e" (UID: "181b6931-2498-4979-a9b1-f6cbc925c78e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.105567 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.105611 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmnjj\" (UniqueName: \"kubernetes.io/projected/181b6931-2498-4979-a9b1-f6cbc925c78e-kube-api-access-rmnjj\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.105626 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.105637 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/181b6931-2498-4979-a9b1-f6cbc925c78e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.172701 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-b02d-account-create-update-vkxp4" event={"ID":"6c983cab-0958-4f0d-964c-1a188cad056e","Type":"ContainerStarted","Data":"ce7d1a473b3058d85746221f40269742773a94dfe0266450ac3b299f125b1993"} Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.172746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-b02d-account-create-update-vkxp4" event={"ID":"6c983cab-0958-4f0d-964c-1a188cad056e","Type":"ContainerStarted","Data":"4883718b75355f2f8382f0615481f1dd1d8126faf5abc48b9715258859bb71be"} Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.176728 4744 generic.go:334] "Generic (PLEG): container finished" podID="2d75d787-92a2-4fe3-8783-73d3029693a9" containerID="27dd7ce83dae3b277bcb9185ef4f981f9f854b19b7c8820266053217f5e5e03a" exitCode=0 Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.176791 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-f4krx" event={"ID":"2d75d787-92a2-4fe3-8783-73d3029693a9","Type":"ContainerDied","Data":"27dd7ce83dae3b277bcb9185ef4f981f9f854b19b7c8820266053217f5e5e03a"} Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.176819 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-f4krx" event={"ID":"2d75d787-92a2-4fe3-8783-73d3029693a9","Type":"ContainerStarted","Data":"45fe4e7cac4dc4a8c2804b58e60d92875a9e50d2d67d84ed33f53f7e91802428"} Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.179812 4744 generic.go:334] "Generic (PLEG): container finished" podID="181b6931-2498-4979-a9b1-f6cbc925c78e" containerID="ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b" exitCode=137 Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.179905 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.180001 4744 scope.go:117] "RemoveContainer" containerID="ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.193130 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-b02d-account-create-update-vkxp4" podStartSLOduration=2.193111848 podStartE2EDuration="2.193111848s" podCreationTimestamp="2025-12-01 09:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:06:03.19284172 +0000 UTC m=+2895.181899651" watchObservedRunningTime="2025-12-01 09:06:03.193111848 +0000 UTC m=+2895.182169779" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.218865 4744 scope.go:117] "RemoveContainer" containerID="ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b" Dec 01 09:06:03 crc kubenswrapper[4744]: E1201 09:06:03.219584 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b\": container with ID starting with ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b not found: ID does not exist" containerID="ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.219628 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b"} err="failed to get container status \"ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b\": rpc error: code = NotFound desc = could not find container \"ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b\": container with ID starting with ba6f483e04a10b50b382c78599c4d0d349ef12bbf707cfc54d76a3da17c2d17b not found: ID does not exist" Dec 01 09:06:03 crc kubenswrapper[4744]: I1201 09:06:03.220050 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="181b6931-2498-4979-a9b1-f6cbc925c78e" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.191830 4744 generic.go:334] "Generic (PLEG): container finished" podID="6c983cab-0958-4f0d-964c-1a188cad056e" containerID="ce7d1a473b3058d85746221f40269742773a94dfe0266450ac3b299f125b1993" exitCode=0 Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.191886 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-b02d-account-create-update-vkxp4" event={"ID":"6c983cab-0958-4f0d-964c-1a188cad056e","Type":"ContainerDied","Data":"ce7d1a473b3058d85746221f40269742773a94dfe0266450ac3b299f125b1993"} Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.300183 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="181b6931-2498-4979-a9b1-f6cbc925c78e" path="/var/lib/kubelet/pods/181b6931-2498-4979-a9b1-f6cbc925c78e/volumes" Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.538740 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.635815 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d75d787-92a2-4fe3-8783-73d3029693a9-operator-scripts\") pod \"2d75d787-92a2-4fe3-8783-73d3029693a9\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.635981 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9pck\" (UniqueName: \"kubernetes.io/projected/2d75d787-92a2-4fe3-8783-73d3029693a9-kube-api-access-t9pck\") pod \"2d75d787-92a2-4fe3-8783-73d3029693a9\" (UID: \"2d75d787-92a2-4fe3-8783-73d3029693a9\") " Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.636887 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d75d787-92a2-4fe3-8783-73d3029693a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d75d787-92a2-4fe3-8783-73d3029693a9" (UID: "2d75d787-92a2-4fe3-8783-73d3029693a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.640764 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d75d787-92a2-4fe3-8783-73d3029693a9-kube-api-access-t9pck" (OuterVolumeSpecName: "kube-api-access-t9pck") pod "2d75d787-92a2-4fe3-8783-73d3029693a9" (UID: "2d75d787-92a2-4fe3-8783-73d3029693a9"). InnerVolumeSpecName "kube-api-access-t9pck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.739050 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9pck\" (UniqueName: \"kubernetes.io/projected/2d75d787-92a2-4fe3-8783-73d3029693a9-kube-api-access-t9pck\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:04 crc kubenswrapper[4744]: I1201 09:06:04.739099 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d75d787-92a2-4fe3-8783-73d3029693a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.204709 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-f4krx" event={"ID":"2d75d787-92a2-4fe3-8783-73d3029693a9","Type":"ContainerDied","Data":"45fe4e7cac4dc4a8c2804b58e60d92875a9e50d2d67d84ed33f53f7e91802428"} Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.205149 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45fe4e7cac4dc4a8c2804b58e60d92875a9e50d2d67d84ed33f53f7e91802428" Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.204740 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-f4krx" Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.588982 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.757437 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsbng\" (UniqueName: \"kubernetes.io/projected/6c983cab-0958-4f0d-964c-1a188cad056e-kube-api-access-nsbng\") pod \"6c983cab-0958-4f0d-964c-1a188cad056e\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.757723 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c983cab-0958-4f0d-964c-1a188cad056e-operator-scripts\") pod \"6c983cab-0958-4f0d-964c-1a188cad056e\" (UID: \"6c983cab-0958-4f0d-964c-1a188cad056e\") " Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.758426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c983cab-0958-4f0d-964c-1a188cad056e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c983cab-0958-4f0d-964c-1a188cad056e" (UID: "6c983cab-0958-4f0d-964c-1a188cad056e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.768647 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c983cab-0958-4f0d-964c-1a188cad056e-kube-api-access-nsbng" (OuterVolumeSpecName: "kube-api-access-nsbng") pod "6c983cab-0958-4f0d-964c-1a188cad056e" (UID: "6c983cab-0958-4f0d-964c-1a188cad056e"). InnerVolumeSpecName "kube-api-access-nsbng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.859615 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsbng\" (UniqueName: \"kubernetes.io/projected/6c983cab-0958-4f0d-964c-1a188cad056e-kube-api-access-nsbng\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:05 crc kubenswrapper[4744]: I1201 09:06:05.859886 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c983cab-0958-4f0d-964c-1a188cad056e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:06 crc kubenswrapper[4744]: I1201 09:06:06.230937 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-b02d-account-create-update-vkxp4" event={"ID":"6c983cab-0958-4f0d-964c-1a188cad056e","Type":"ContainerDied","Data":"4883718b75355f2f8382f0615481f1dd1d8126faf5abc48b9715258859bb71be"} Dec 01 09:06:06 crc kubenswrapper[4744]: I1201 09:06:06.230981 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4883718b75355f2f8382f0615481f1dd1d8126faf5abc48b9715258859bb71be" Dec 01 09:06:06 crc kubenswrapper[4744]: I1201 09:06:06.231016 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-b02d-account-create-update-vkxp4" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.320616 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-tlszx"] Dec 01 09:06:07 crc kubenswrapper[4744]: E1201 09:06:07.321064 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d75d787-92a2-4fe3-8783-73d3029693a9" containerName="mariadb-database-create" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.321083 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d75d787-92a2-4fe3-8783-73d3029693a9" containerName="mariadb-database-create" Dec 01 09:06:07 crc kubenswrapper[4744]: E1201 09:06:07.321129 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c983cab-0958-4f0d-964c-1a188cad056e" containerName="mariadb-account-create-update" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.321139 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c983cab-0958-4f0d-964c-1a188cad056e" containerName="mariadb-account-create-update" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.321377 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c983cab-0958-4f0d-964c-1a188cad056e" containerName="mariadb-account-create-update" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.321390 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d75d787-92a2-4fe3-8783-73d3029693a9" containerName="mariadb-database-create" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.322329 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.325740 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.325922 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.326050 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.326891 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-h64d7" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.377814 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-tlszx"] Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.413618 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r94cs\" (UniqueName: \"kubernetes.io/projected/bf79a6a4-97c1-4547-9d12-05c8d57534a8-kube-api-access-r94cs\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.413700 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-scripts\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.413852 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-combined-ca-bundle\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.413950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-config-data\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.516046 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-combined-ca-bundle\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.516369 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-config-data\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.516554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r94cs\" (UniqueName: \"kubernetes.io/projected/bf79a6a4-97c1-4547-9d12-05c8d57534a8-kube-api-access-r94cs\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.516686 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-scripts\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.521805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-combined-ca-bundle\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.523001 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-scripts\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.533141 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r94cs\" (UniqueName: \"kubernetes.io/projected/bf79a6a4-97c1-4547-9d12-05c8d57534a8-kube-api-access-r94cs\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.534142 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-config-data\") pod \"aodh-db-sync-tlszx\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:07 crc kubenswrapper[4744]: I1201 09:06:07.638508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:08 crc kubenswrapper[4744]: I1201 09:06:08.100024 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:06:08 crc kubenswrapper[4744]: I1201 09:06:08.105767 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-tlszx"] Dec 01 09:06:08 crc kubenswrapper[4744]: I1201 09:06:08.257715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tlszx" event={"ID":"bf79a6a4-97c1-4547-9d12-05c8d57534a8","Type":"ContainerStarted","Data":"0ac43b6d619d2eb4428a2f26d455729f2b7a10bb2789da0865540803b1dd7c9f"} Dec 01 09:06:12 crc kubenswrapper[4744]: I1201 09:06:12.325159 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tlszx" event={"ID":"bf79a6a4-97c1-4547-9d12-05c8d57534a8","Type":"ContainerStarted","Data":"c9641cc77d338fe39b40598726b700dc87cb789e0818a0b286387c7538d98e1c"} Dec 01 09:06:12 crc kubenswrapper[4744]: I1201 09:06:12.356679 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-tlszx" podStartSLOduration=1.476210585 podStartE2EDuration="5.356656968s" podCreationTimestamp="2025-12-01 09:06:07 +0000 UTC" firstStartedPulling="2025-12-01 09:06:08.099634669 +0000 UTC m=+2900.088692600" lastFinishedPulling="2025-12-01 09:06:11.980081072 +0000 UTC m=+2903.969138983" observedRunningTime="2025-12-01 09:06:12.35359555 +0000 UTC m=+2904.342653511" watchObservedRunningTime="2025-12-01 09:06:12.356656968 +0000 UTC m=+2904.345714899" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.773112 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgl"] Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.776530 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.788309 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgl"] Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.858807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-utilities\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.859387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-catalog-content\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.859515 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68lv7\" (UniqueName: \"kubernetes.io/projected/054cf1b9-c9ef-4ec8-94d0-b88b90109442-kube-api-access-68lv7\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.962010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-catalog-content\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.962115 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68lv7\" (UniqueName: \"kubernetes.io/projected/054cf1b9-c9ef-4ec8-94d0-b88b90109442-kube-api-access-68lv7\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.962474 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-utilities\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.962608 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-catalog-content\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:13 crc kubenswrapper[4744]: I1201 09:06:13.963038 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-utilities\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:14 crc kubenswrapper[4744]: I1201 09:06:14.001027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68lv7\" (UniqueName: \"kubernetes.io/projected/054cf1b9-c9ef-4ec8-94d0-b88b90109442-kube-api-access-68lv7\") pod \"redhat-marketplace-nztgl\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:14 crc kubenswrapper[4744]: I1201 09:06:14.109620 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:14 crc kubenswrapper[4744]: I1201 09:06:14.380708 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgl"] Dec 01 09:06:14 crc kubenswrapper[4744]: W1201 09:06:14.385104 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod054cf1b9_c9ef_4ec8_94d0_b88b90109442.slice/crio-d96af258d1c45a6b5fcd5c377c69660ec80ea92397e0eeae9619cf49d25a86f1 WatchSource:0}: Error finding container d96af258d1c45a6b5fcd5c377c69660ec80ea92397e0eeae9619cf49d25a86f1: Status 404 returned error can't find the container with id d96af258d1c45a6b5fcd5c377c69660ec80ea92397e0eeae9619cf49d25a86f1 Dec 01 09:06:15 crc kubenswrapper[4744]: I1201 09:06:15.354199 4744 generic.go:334] "Generic (PLEG): container finished" podID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerID="17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186" exitCode=0 Dec 01 09:06:15 crc kubenswrapper[4744]: I1201 09:06:15.354394 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgl" event={"ID":"054cf1b9-c9ef-4ec8-94d0-b88b90109442","Type":"ContainerDied","Data":"17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186"} Dec 01 09:06:15 crc kubenswrapper[4744]: I1201 09:06:15.354669 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgl" event={"ID":"054cf1b9-c9ef-4ec8-94d0-b88b90109442","Type":"ContainerStarted","Data":"d96af258d1c45a6b5fcd5c377c69660ec80ea92397e0eeae9619cf49d25a86f1"} Dec 01 09:06:15 crc kubenswrapper[4744]: I1201 09:06:15.356972 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf79a6a4-97c1-4547-9d12-05c8d57534a8" containerID="c9641cc77d338fe39b40598726b700dc87cb789e0818a0b286387c7538d98e1c" exitCode=0 Dec 01 09:06:15 crc kubenswrapper[4744]: I1201 09:06:15.357020 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tlszx" event={"ID":"bf79a6a4-97c1-4547-9d12-05c8d57534a8","Type":"ContainerDied","Data":"c9641cc77d338fe39b40598726b700dc87cb789e0818a0b286387c7538d98e1c"} Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.369383 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgl" event={"ID":"054cf1b9-c9ef-4ec8-94d0-b88b90109442","Type":"ContainerStarted","Data":"5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790"} Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.737864 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.827150 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-combined-ca-bundle\") pod \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.827736 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-scripts\") pod \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.827790 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-config-data\") pod \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.827838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r94cs\" (UniqueName: \"kubernetes.io/projected/bf79a6a4-97c1-4547-9d12-05c8d57534a8-kube-api-access-r94cs\") pod \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\" (UID: \"bf79a6a4-97c1-4547-9d12-05c8d57534a8\") " Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.838020 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-scripts" (OuterVolumeSpecName: "scripts") pod "bf79a6a4-97c1-4547-9d12-05c8d57534a8" (UID: "bf79a6a4-97c1-4547-9d12-05c8d57534a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.840991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf79a6a4-97c1-4547-9d12-05c8d57534a8-kube-api-access-r94cs" (OuterVolumeSpecName: "kube-api-access-r94cs") pod "bf79a6a4-97c1-4547-9d12-05c8d57534a8" (UID: "bf79a6a4-97c1-4547-9d12-05c8d57534a8"). InnerVolumeSpecName "kube-api-access-r94cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.866497 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-config-data" (OuterVolumeSpecName: "config-data") pod "bf79a6a4-97c1-4547-9d12-05c8d57534a8" (UID: "bf79a6a4-97c1-4547-9d12-05c8d57534a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.868869 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf79a6a4-97c1-4547-9d12-05c8d57534a8" (UID: "bf79a6a4-97c1-4547-9d12-05c8d57534a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.930021 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.930060 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.930094 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf79a6a4-97c1-4547-9d12-05c8d57534a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:16 crc kubenswrapper[4744]: I1201 09:06:16.930106 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r94cs\" (UniqueName: \"kubernetes.io/projected/bf79a6a4-97c1-4547-9d12-05c8d57534a8-kube-api-access-r94cs\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:17 crc kubenswrapper[4744]: I1201 09:06:17.381291 4744 generic.go:334] "Generic (PLEG): container finished" podID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerID="5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790" exitCode=0 Dec 01 09:06:17 crc kubenswrapper[4744]: I1201 09:06:17.381371 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgl" event={"ID":"054cf1b9-c9ef-4ec8-94d0-b88b90109442","Type":"ContainerDied","Data":"5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790"} Dec 01 09:06:17 crc kubenswrapper[4744]: I1201 09:06:17.384073 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tlszx" event={"ID":"bf79a6a4-97c1-4547-9d12-05c8d57534a8","Type":"ContainerDied","Data":"0ac43b6d619d2eb4428a2f26d455729f2b7a10bb2789da0865540803b1dd7c9f"} Dec 01 09:06:17 crc kubenswrapper[4744]: I1201 09:06:17.384108 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tlszx" Dec 01 09:06:17 crc kubenswrapper[4744]: I1201 09:06:17.384108 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ac43b6d619d2eb4428a2f26d455729f2b7a10bb2789da0865540803b1dd7c9f" Dec 01 09:06:18 crc kubenswrapper[4744]: I1201 09:06:18.398564 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgl" event={"ID":"054cf1b9-c9ef-4ec8-94d0-b88b90109442","Type":"ContainerStarted","Data":"01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7"} Dec 01 09:06:18 crc kubenswrapper[4744]: I1201 09:06:18.423270 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nztgl" podStartSLOduration=2.888818506 podStartE2EDuration="5.423252648s" podCreationTimestamp="2025-12-01 09:06:13 +0000 UTC" firstStartedPulling="2025-12-01 09:06:15.356857404 +0000 UTC m=+2907.345915335" lastFinishedPulling="2025-12-01 09:06:17.891291556 +0000 UTC m=+2909.880349477" observedRunningTime="2025-12-01 09:06:18.41672206 +0000 UTC m=+2910.405779991" watchObservedRunningTime="2025-12-01 09:06:18.423252648 +0000 UTC m=+2910.412310569" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.933940 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 09:06:21 crc kubenswrapper[4744]: E1201 09:06:21.935027 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf79a6a4-97c1-4547-9d12-05c8d57534a8" containerName="aodh-db-sync" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.935046 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf79a6a4-97c1-4547-9d12-05c8d57534a8" containerName="aodh-db-sync" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.935269 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf79a6a4-97c1-4547-9d12-05c8d57534a8" containerName="aodh-db-sync" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.938357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.940604 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.940820 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-h64d7" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.942380 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 09:06:21 crc kubenswrapper[4744]: I1201 09:06:21.947156 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.053535 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-scripts\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.053857 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.053888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9jrc\" (UniqueName: \"kubernetes.io/projected/e439e1a0-24b7-4622-a5ba-1f5d940947f4-kube-api-access-v9jrc\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.053952 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-config-data\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.155303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-scripts\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.155359 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.155390 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9jrc\" (UniqueName: \"kubernetes.io/projected/e439e1a0-24b7-4622-a5ba-1f5d940947f4-kube-api-access-v9jrc\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.155468 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-config-data\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.162360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.164139 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-scripts\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.166050 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-config-data\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.174222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9jrc\" (UniqueName: \"kubernetes.io/projected/e439e1a0-24b7-4622-a5ba-1f5d940947f4-kube-api-access-v9jrc\") pod \"aodh-0\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.256183 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:06:22 crc kubenswrapper[4744]: I1201 09:06:22.735041 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:06:22 crc kubenswrapper[4744]: W1201 09:06:22.738765 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode439e1a0_24b7_4622_a5ba_1f5d940947f4.slice/crio-a873a196cdefa31503f369b1c51c1affebc18fc4e724c7ff98d4d3de8921b13f WatchSource:0}: Error finding container a873a196cdefa31503f369b1c51c1affebc18fc4e724c7ff98d4d3de8921b13f: Status 404 returned error can't find the container with id a873a196cdefa31503f369b1c51c1affebc18fc4e724c7ff98d4d3de8921b13f Dec 01 09:06:23 crc kubenswrapper[4744]: I1201 09:06:23.462957 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerStarted","Data":"a873a196cdefa31503f369b1c51c1affebc18fc4e724c7ff98d4d3de8921b13f"} Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.110550 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.110928 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.171761 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.172113 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="proxy-httpd" containerID="cri-o://a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06" gracePeriod=30 Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.172111 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-central-agent" containerID="cri-o://eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed" gracePeriod=30 Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.172159 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="sg-core" containerID="cri-o://91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd" gracePeriod=30 Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.172212 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-notification-agent" containerID="cri-o://2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d" gracePeriod=30 Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.184003 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.472290 4744 generic.go:334] "Generic (PLEG): container finished" podID="340586bd-ec9d-4a58-a46d-1d043557b994" containerID="a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06" exitCode=0 Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.472620 4744 generic.go:334] "Generic (PLEG): container finished" podID="340586bd-ec9d-4a58-a46d-1d043557b994" containerID="91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd" exitCode=2 Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.472371 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerDied","Data":"a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06"} Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.472736 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerDied","Data":"91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd"} Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.474399 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerStarted","Data":"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9"} Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.528855 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.583874 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgl"] Dec 01 09:06:24 crc kubenswrapper[4744]: I1201 09:06:24.859794 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:06:25 crc kubenswrapper[4744]: I1201 09:06:25.485831 4744 generic.go:334] "Generic (PLEG): container finished" podID="340586bd-ec9d-4a58-a46d-1d043557b994" containerID="eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed" exitCode=0 Dec 01 09:06:25 crc kubenswrapper[4744]: I1201 09:06:25.486007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerDied","Data":"eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed"} Dec 01 09:06:26 crc kubenswrapper[4744]: I1201 09:06:26.497162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerStarted","Data":"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b"} Dec 01 09:06:26 crc kubenswrapper[4744]: I1201 09:06:26.497331 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nztgl" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="registry-server" containerID="cri-o://01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7" gracePeriod=2 Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.045535 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.168182 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-catalog-content\") pod \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.168228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-utilities\") pod \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.168311 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68lv7\" (UniqueName: \"kubernetes.io/projected/054cf1b9-c9ef-4ec8-94d0-b88b90109442-kube-api-access-68lv7\") pod \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\" (UID: \"054cf1b9-c9ef-4ec8-94d0-b88b90109442\") " Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.169947 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-utilities" (OuterVolumeSpecName: "utilities") pod "054cf1b9-c9ef-4ec8-94d0-b88b90109442" (UID: "054cf1b9-c9ef-4ec8-94d0-b88b90109442"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.188813 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/054cf1b9-c9ef-4ec8-94d0-b88b90109442-kube-api-access-68lv7" (OuterVolumeSpecName: "kube-api-access-68lv7") pod "054cf1b9-c9ef-4ec8-94d0-b88b90109442" (UID: "054cf1b9-c9ef-4ec8-94d0-b88b90109442"). InnerVolumeSpecName "kube-api-access-68lv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.190055 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "054cf1b9-c9ef-4ec8-94d0-b88b90109442" (UID: "054cf1b9-c9ef-4ec8-94d0-b88b90109442"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.270719 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.270763 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/054cf1b9-c9ef-4ec8-94d0-b88b90109442-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.270775 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68lv7\" (UniqueName: \"kubernetes.io/projected/054cf1b9-c9ef-4ec8-94d0-b88b90109442-kube-api-access-68lv7\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.512087 4744 generic.go:334] "Generic (PLEG): container finished" podID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerID="01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7" exitCode=0 Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.512630 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgl" event={"ID":"054cf1b9-c9ef-4ec8-94d0-b88b90109442","Type":"ContainerDied","Data":"01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7"} Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.512660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nztgl" event={"ID":"054cf1b9-c9ef-4ec8-94d0-b88b90109442","Type":"ContainerDied","Data":"d96af258d1c45a6b5fcd5c377c69660ec80ea92397e0eeae9619cf49d25a86f1"} Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.512703 4744 scope.go:117] "RemoveContainer" containerID="01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.512871 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nztgl" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.554250 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgl"] Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.556487 4744 scope.go:117] "RemoveContainer" containerID="5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.567374 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nztgl"] Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.576649 4744 scope.go:117] "RemoveContainer" containerID="17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.623679 4744 scope.go:117] "RemoveContainer" containerID="01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7" Dec 01 09:06:27 crc kubenswrapper[4744]: E1201 09:06:27.624784 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7\": container with ID starting with 01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7 not found: ID does not exist" containerID="01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.624824 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7"} err="failed to get container status \"01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7\": rpc error: code = NotFound desc = could not find container \"01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7\": container with ID starting with 01e16a7b8f3d18dc1ad1d273e6c6f688a97deb470c784df721cee0e26cef09f7 not found: ID does not exist" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.624850 4744 scope.go:117] "RemoveContainer" containerID="5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790" Dec 01 09:06:27 crc kubenswrapper[4744]: E1201 09:06:27.625183 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790\": container with ID starting with 5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790 not found: ID does not exist" containerID="5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.625222 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790"} err="failed to get container status \"5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790\": rpc error: code = NotFound desc = could not find container \"5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790\": container with ID starting with 5771b9e654d1c9e0831be36f876dbd0753c28f08d327b2cf5e4009d1255d3790 not found: ID does not exist" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.625249 4744 scope.go:117] "RemoveContainer" containerID="17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186" Dec 01 09:06:27 crc kubenswrapper[4744]: E1201 09:06:27.625695 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186\": container with ID starting with 17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186 not found: ID does not exist" containerID="17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186" Dec 01 09:06:27 crc kubenswrapper[4744]: I1201 09:06:27.625716 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186"} err="failed to get container status \"17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186\": rpc error: code = NotFound desc = could not find container \"17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186\": container with ID starting with 17e93c86741e9e7afd1daa2f2f36c264014d35790f7bf829a5884fa5ef4ef186 not found: ID does not exist" Dec 01 09:06:28 crc kubenswrapper[4744]: I1201 09:06:28.295589 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" path="/var/lib/kubelet/pods/054cf1b9-c9ef-4ec8-94d0-b88b90109442/volumes" Dec 01 09:06:28 crc kubenswrapper[4744]: I1201 09:06:28.525106 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerStarted","Data":"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da"} Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.323542 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.457549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-sg-core-conf-yaml\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.457893 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-combined-ca-bundle\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-config-data\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-ceilometer-tls-certs\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-log-httpd\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458195 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-run-httpd\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rklpx\" (UniqueName: \"kubernetes.io/projected/340586bd-ec9d-4a58-a46d-1d043557b994-kube-api-access-rklpx\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458612 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458693 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-scripts\") pod \"340586bd-ec9d-4a58-a46d-1d043557b994\" (UID: \"340586bd-ec9d-4a58-a46d-1d043557b994\") " Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.458728 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.459219 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.459246 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/340586bd-ec9d-4a58-a46d-1d043557b994-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.461882 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/340586bd-ec9d-4a58-a46d-1d043557b994-kube-api-access-rklpx" (OuterVolumeSpecName: "kube-api-access-rklpx") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "kube-api-access-rklpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.461867 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-scripts" (OuterVolumeSpecName: "scripts") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.487114 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.508069 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.534191 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.561253 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.561283 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rklpx\" (UniqueName: \"kubernetes.io/projected/340586bd-ec9d-4a58-a46d-1d043557b994-kube-api-access-rklpx\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.561295 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.561303 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.561312 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.562923 4744 generic.go:334] "Generic (PLEG): container finished" podID="340586bd-ec9d-4a58-a46d-1d043557b994" containerID="2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d" exitCode=0 Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.562982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerDied","Data":"2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d"} Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.563038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"340586bd-ec9d-4a58-a46d-1d043557b994","Type":"ContainerDied","Data":"75ac7023db2237a9400f179e7bf2ae916b06dcca79cba17a9a4fcd2ca5fd3bc0"} Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.563057 4744 scope.go:117] "RemoveContainer" containerID="a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.562984 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.566620 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerStarted","Data":"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a"} Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.566772 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-api" containerID="cri-o://aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9" gracePeriod=30 Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.566855 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-listener" containerID="cri-o://f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a" gracePeriod=30 Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.566912 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-notifier" containerID="cri-o://2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da" gracePeriod=30 Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.566958 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-evaluator" containerID="cri-o://0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b" gracePeriod=30 Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.577923 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-config-data" (OuterVolumeSpecName: "config-data") pod "340586bd-ec9d-4a58-a46d-1d043557b994" (UID: "340586bd-ec9d-4a58-a46d-1d043557b994"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.585382 4744 scope.go:117] "RemoveContainer" containerID="91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.610085 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.270461138 podStartE2EDuration="10.610061477s" podCreationTimestamp="2025-12-01 09:06:21 +0000 UTC" firstStartedPulling="2025-12-01 09:06:22.742144741 +0000 UTC m=+2914.731202662" lastFinishedPulling="2025-12-01 09:06:31.08174508 +0000 UTC m=+2923.070803001" observedRunningTime="2025-12-01 09:06:31.600877863 +0000 UTC m=+2923.589935804" watchObservedRunningTime="2025-12-01 09:06:31.610061477 +0000 UTC m=+2923.599119398" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.617660 4744 scope.go:117] "RemoveContainer" containerID="2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.640132 4744 scope.go:117] "RemoveContainer" containerID="eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.663181 4744 scope.go:117] "RemoveContainer" containerID="a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.663855 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06\": container with ID starting with a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06 not found: ID does not exist" containerID="a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.663891 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06"} err="failed to get container status \"a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06\": rpc error: code = NotFound desc = could not find container \"a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06\": container with ID starting with a1317800858bfa987e01836ef49e18abf06adedcc5e09d66589c65c46d0a0f06 not found: ID does not exist" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.663947 4744 scope.go:117] "RemoveContainer" containerID="91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.663962 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340586bd-ec9d-4a58-a46d-1d043557b994-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.664226 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd\": container with ID starting with 91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd not found: ID does not exist" containerID="91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.664247 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd"} err="failed to get container status \"91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd\": rpc error: code = NotFound desc = could not find container \"91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd\": container with ID starting with 91ba6a7854efb67019a94352d99d11dee1cb619c653c7d7f7fde72b1a4b871cd not found: ID does not exist" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.664277 4744 scope.go:117] "RemoveContainer" containerID="2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.664567 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d\": container with ID starting with 2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d not found: ID does not exist" containerID="2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.664589 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d"} err="failed to get container status \"2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d\": rpc error: code = NotFound desc = could not find container \"2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d\": container with ID starting with 2fe2c81de05be56fa6380464d0d23dbb2078118c92383edc16e9e22d8ce7ce9d not found: ID does not exist" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.664623 4744 scope.go:117] "RemoveContainer" containerID="eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.665535 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed\": container with ID starting with eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed not found: ID does not exist" containerID="eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.665574 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed"} err="failed to get container status \"eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed\": rpc error: code = NotFound desc = could not find container \"eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed\": container with ID starting with eadd06b32f5540d0ea5c29b08c27b7c8ada6f202733d69af40926bb5df6280ed not found: ID does not exist" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.899797 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.908759 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926266 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.926715 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="sg-core" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926733 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="sg-core" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.926766 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-central-agent" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926773 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-central-agent" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.926787 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="extract-content" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926794 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="extract-content" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.926807 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="proxy-httpd" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926812 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="proxy-httpd" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.926826 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="extract-utilities" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926833 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="extract-utilities" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.926848 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="registry-server" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926853 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="registry-server" Dec 01 09:06:31 crc kubenswrapper[4744]: E1201 09:06:31.926861 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-notification-agent" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.926867 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-notification-agent" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.927023 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-central-agent" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.927036 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="ceilometer-notification-agent" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.927047 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="sg-core" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.927060 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" containerName="proxy-httpd" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.927074 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="054cf1b9-c9ef-4ec8-94d0-b88b90109442" containerName="registry-server" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.929002 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.931585 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.932025 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.932183 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 09:06:31 crc kubenswrapper[4744]: I1201 09:06:31.947273 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072612 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-scripts\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072641 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-config-data\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072716 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nhwq\" (UniqueName: \"kubernetes.io/projected/999f566b-3f04-4cfd-9ff8-be798f51635a-kube-api-access-7nhwq\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/999f566b-3f04-4cfd-9ff8-be798f51635a-run-httpd\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072828 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/999f566b-3f04-4cfd-9ff8-be798f51635a-log-httpd\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.072884 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/999f566b-3f04-4cfd-9ff8-be798f51635a-run-httpd\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174656 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174684 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/999f566b-3f04-4cfd-9ff8-be798f51635a-log-httpd\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174748 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174845 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-scripts\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174896 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-config-data\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.174942 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nhwq\" (UniqueName: \"kubernetes.io/projected/999f566b-3f04-4cfd-9ff8-be798f51635a-kube-api-access-7nhwq\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.176004 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/999f566b-3f04-4cfd-9ff8-be798f51635a-run-httpd\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.177006 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/999f566b-3f04-4cfd-9ff8-be798f51635a-log-httpd\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.181125 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.181133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.182052 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-scripts\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.194510 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.196821 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/999f566b-3f04-4cfd-9ff8-be798f51635a-config-data\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.197275 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nhwq\" (UniqueName: \"kubernetes.io/projected/999f566b-3f04-4cfd-9ff8-be798f51635a-kube-api-access-7nhwq\") pod \"ceilometer-0\" (UID: \"999f566b-3f04-4cfd-9ff8-be798f51635a\") " pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.249176 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.300429 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="340586bd-ec9d-4a58-a46d-1d043557b994" path="/var/lib/kubelet/pods/340586bd-ec9d-4a58-a46d-1d043557b994/volumes" Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.577165 4744 generic.go:334] "Generic (PLEG): container finished" podID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerID="0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b" exitCode=0 Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.577674 4744 generic.go:334] "Generic (PLEG): container finished" podID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerID="aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9" exitCode=0 Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.577270 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerDied","Data":"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b"} Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.577771 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerDied","Data":"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9"} Dec 01 09:06:32 crc kubenswrapper[4744]: I1201 09:06:32.722836 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 09:06:33 crc kubenswrapper[4744]: I1201 09:06:33.589494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"999f566b-3f04-4cfd-9ff8-be798f51635a","Type":"ContainerStarted","Data":"d28fb66cc68c100aae368353ae4aca541b53a09e52cdcf29990be7768d774616"} Dec 01 09:06:34 crc kubenswrapper[4744]: I1201 09:06:34.599129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"999f566b-3f04-4cfd-9ff8-be798f51635a","Type":"ContainerStarted","Data":"b8f7732edb0fff1f27b52a2ac088ff1adda075736be11bd7674f1f28f92cd33a"} Dec 01 09:06:35 crc kubenswrapper[4744]: I1201 09:06:35.610663 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"999f566b-3f04-4cfd-9ff8-be798f51635a","Type":"ContainerStarted","Data":"1888a384f9c74654af0051f9b7687d1ed8ea52552864adfe67f523208dc590eb"} Dec 01 09:06:36 crc kubenswrapper[4744]: I1201 09:06:36.624082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"999f566b-3f04-4cfd-9ff8-be798f51635a","Type":"ContainerStarted","Data":"f127ae0220cc88c29af0d75a11e4b96162d25f369d29ce58000a4fd2d029a5ed"} Dec 01 09:06:39 crc kubenswrapper[4744]: I1201 09:06:39.654378 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"999f566b-3f04-4cfd-9ff8-be798f51635a","Type":"ContainerStarted","Data":"b0b5d49812b431de1a77296e7cd420b9c5ed5d82a4dfcc77091865668e1f3d17"} Dec 01 09:06:39 crc kubenswrapper[4744]: I1201 09:06:39.654914 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 09:06:48 crc kubenswrapper[4744]: I1201 09:06:48.634784 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:06:48 crc kubenswrapper[4744]: I1201 09:06:48.635583 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.257709 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.293762 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=25.509010791 podStartE2EDuration="31.29374228s" podCreationTimestamp="2025-12-01 09:06:31 +0000 UTC" firstStartedPulling="2025-12-01 09:06:32.723325253 +0000 UTC m=+2924.712383174" lastFinishedPulling="2025-12-01 09:06:38.508056742 +0000 UTC m=+2930.497114663" observedRunningTime="2025-12-01 09:06:39.68552944 +0000 UTC m=+2931.674587361" watchObservedRunningTime="2025-12-01 09:07:02.29374228 +0000 UTC m=+2954.282800211" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.623900 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.801039 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-config-data\") pod \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.801084 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-combined-ca-bundle\") pod \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.801255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-scripts\") pod \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.801296 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9jrc\" (UniqueName: \"kubernetes.io/projected/e439e1a0-24b7-4622-a5ba-1f5d940947f4-kube-api-access-v9jrc\") pod \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\" (UID: \"e439e1a0-24b7-4622-a5ba-1f5d940947f4\") " Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.806717 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e439e1a0-24b7-4622-a5ba-1f5d940947f4-kube-api-access-v9jrc" (OuterVolumeSpecName: "kube-api-access-v9jrc") pod "e439e1a0-24b7-4622-a5ba-1f5d940947f4" (UID: "e439e1a0-24b7-4622-a5ba-1f5d940947f4"). InnerVolumeSpecName "kube-api-access-v9jrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.807987 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-scripts" (OuterVolumeSpecName: "scripts") pod "e439e1a0-24b7-4622-a5ba-1f5d940947f4" (UID: "e439e1a0-24b7-4622-a5ba-1f5d940947f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.902707 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e439e1a0-24b7-4622-a5ba-1f5d940947f4" (UID: "e439e1a0-24b7-4622-a5ba-1f5d940947f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.903115 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.903146 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9jrc\" (UniqueName: \"kubernetes.io/projected/e439e1a0-24b7-4622-a5ba-1f5d940947f4-kube-api-access-v9jrc\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.903159 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.904782 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-config-data" (OuterVolumeSpecName: "config-data") pod "e439e1a0-24b7-4622-a5ba-1f5d940947f4" (UID: "e439e1a0-24b7-4622-a5ba-1f5d940947f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.912913 4744 generic.go:334] "Generic (PLEG): container finished" podID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerID="f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a" exitCode=137 Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.912949 4744 generic.go:334] "Generic (PLEG): container finished" podID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerID="2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da" exitCode=137 Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.912965 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.912968 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerDied","Data":"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a"} Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.913002 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerDied","Data":"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da"} Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.913015 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e439e1a0-24b7-4622-a5ba-1f5d940947f4","Type":"ContainerDied","Data":"a873a196cdefa31503f369b1c51c1affebc18fc4e724c7ff98d4d3de8921b13f"} Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.913031 4744 scope.go:117] "RemoveContainer" containerID="f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a" Dec 01 09:07:02 crc kubenswrapper[4744]: I1201 09:07:02.991391 4744 scope.go:117] "RemoveContainer" containerID="2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.004618 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e439e1a0-24b7-4622-a5ba-1f5d940947f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.010918 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.023212 4744 scope.go:117] "RemoveContainer" containerID="0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.037631 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.046521 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.047237 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-evaluator" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047265 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-evaluator" Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.047276 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-notifier" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047285 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-notifier" Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.047299 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-api" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047305 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-api" Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.047318 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-listener" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047324 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-listener" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047552 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-listener" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047572 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-api" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047584 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-evaluator" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.047602 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" containerName="aodh-notifier" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.049750 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.052219 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.052507 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.052801 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.052979 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-h64d7" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.053090 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.071259 4744 scope.go:117] "RemoveContainer" containerID="aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.084684 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.096835 4744 scope.go:117] "RemoveContainer" containerID="f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a" Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.097327 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a\": container with ID starting with f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a not found: ID does not exist" containerID="f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.097374 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a"} err="failed to get container status \"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a\": rpc error: code = NotFound desc = could not find container \"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a\": container with ID starting with f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.097510 4744 scope.go:117] "RemoveContainer" containerID="2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da" Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.098073 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da\": container with ID starting with 2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da not found: ID does not exist" containerID="2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.098108 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da"} err="failed to get container status \"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da\": rpc error: code = NotFound desc = could not find container \"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da\": container with ID starting with 2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.098129 4744 scope.go:117] "RemoveContainer" containerID="0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b" Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.098679 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b\": container with ID starting with 0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b not found: ID does not exist" containerID="0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.098709 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b"} err="failed to get container status \"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b\": rpc error: code = NotFound desc = could not find container \"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b\": container with ID starting with 0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.098743 4744 scope.go:117] "RemoveContainer" containerID="aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9" Dec 01 09:07:03 crc kubenswrapper[4744]: E1201 09:07:03.099072 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9\": container with ID starting with aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9 not found: ID does not exist" containerID="aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.099110 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9"} err="failed to get container status \"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9\": rpc error: code = NotFound desc = could not find container \"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9\": container with ID starting with aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9 not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.099135 4744 scope.go:117] "RemoveContainer" containerID="f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.099385 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a"} err="failed to get container status \"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a\": rpc error: code = NotFound desc = could not find container \"f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a\": container with ID starting with f09cf1a4d10bce62652dc077f54482ff0320d73a1c3ce9747af4bc5f19910b3a not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.099441 4744 scope.go:117] "RemoveContainer" containerID="2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.099838 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da"} err="failed to get container status \"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da\": rpc error: code = NotFound desc = could not find container \"2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da\": container with ID starting with 2c5a51ddde454d635d153b38467a4cdbbfda4590792a15ecd519706deb75a0da not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.099862 4744 scope.go:117] "RemoveContainer" containerID="0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.100220 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b"} err="failed to get container status \"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b\": rpc error: code = NotFound desc = could not find container \"0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b\": container with ID starting with 0aeeacc581af093a74c366dd044344b307176b921a33e2297d2b7c31af6c533b not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.100256 4744 scope.go:117] "RemoveContainer" containerID="aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.100571 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9"} err="failed to get container status \"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9\": rpc error: code = NotFound desc = could not find container \"aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9\": container with ID starting with aa5450365ddd8b3856895701989b40f23cf8bb11aa3fac6a738df9c854f1b8e9 not found: ID does not exist" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.107143 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.107499 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-scripts\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.107540 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-public-tls-certs\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.107605 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qjpt\" (UniqueName: \"kubernetes.io/projected/ebb6d047-293a-42e1-9dd2-c35a736f93ea-kube-api-access-4qjpt\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.107652 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-config-data\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.107672 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-internal-tls-certs\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.208571 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.208682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-scripts\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.208704 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-public-tls-certs\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.208728 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qjpt\" (UniqueName: \"kubernetes.io/projected/ebb6d047-293a-42e1-9dd2-c35a736f93ea-kube-api-access-4qjpt\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.208764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-config-data\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.208781 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-internal-tls-certs\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.212725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-scripts\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.212914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-internal-tls-certs\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.213142 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-public-tls-certs\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.214841 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-config-data\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.220885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.230446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qjpt\" (UniqueName: \"kubernetes.io/projected/ebb6d047-293a-42e1-9dd2-c35a736f93ea-kube-api-access-4qjpt\") pod \"aodh-0\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.367295 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.820432 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:07:03 crc kubenswrapper[4744]: W1201 09:07:03.821652 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebb6d047_293a_42e1_9dd2_c35a736f93ea.slice/crio-225beba64fdafff926b32c013efeae9d35c7fae17ae501f006fa45ef50625cbe WatchSource:0}: Error finding container 225beba64fdafff926b32c013efeae9d35c7fae17ae501f006fa45ef50625cbe: Status 404 returned error can't find the container with id 225beba64fdafff926b32c013efeae9d35c7fae17ae501f006fa45ef50625cbe Dec 01 09:07:03 crc kubenswrapper[4744]: I1201 09:07:03.925885 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerStarted","Data":"225beba64fdafff926b32c013efeae9d35c7fae17ae501f006fa45ef50625cbe"} Dec 01 09:07:04 crc kubenswrapper[4744]: I1201 09:07:04.296250 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e439e1a0-24b7-4622-a5ba-1f5d940947f4" path="/var/lib/kubelet/pods/e439e1a0-24b7-4622-a5ba-1f5d940947f4/volumes" Dec 01 09:07:04 crc kubenswrapper[4744]: I1201 09:07:04.937178 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerStarted","Data":"3073308ea771cc8aa2f12ffa6d73e302ce7684c382d0bf167d592bb0b9ff25ee"} Dec 01 09:07:05 crc kubenswrapper[4744]: I1201 09:07:05.950920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerStarted","Data":"8a8dd1d6f9777ebaa902f10e5b908ba2c65e7651dfcc963db3021e84fe0c725a"} Dec 01 09:07:06 crc kubenswrapper[4744]: I1201 09:07:06.962742 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerStarted","Data":"776d6f0bac2b8d73de7f32ce04a1bec2e5dc5d95d742941e6d3a5b548818c1b0"} Dec 01 09:07:07 crc kubenswrapper[4744]: I1201 09:07:07.973955 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerStarted","Data":"a19b8319fa651e231d9ad0039d9b673631d193dec59c60c3c894d33217456a0c"} Dec 01 09:07:18 crc kubenswrapper[4744]: I1201 09:07:18.634208 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:07:18 crc kubenswrapper[4744]: I1201 09:07:18.634883 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:07:48 crc kubenswrapper[4744]: I1201 09:07:48.634946 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:07:48 crc kubenswrapper[4744]: I1201 09:07:48.635698 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:07:48 crc kubenswrapper[4744]: I1201 09:07:48.635766 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 09:07:48 crc kubenswrapper[4744]: I1201 09:07:48.636837 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:07:48 crc kubenswrapper[4744]: I1201 09:07:48.636928 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" gracePeriod=600 Dec 01 09:07:48 crc kubenswrapper[4744]: E1201 09:07:48.776592 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:07:49 crc kubenswrapper[4744]: I1201 09:07:49.390790 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" exitCode=0 Dec 01 09:07:49 crc kubenswrapper[4744]: I1201 09:07:49.390846 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471"} Dec 01 09:07:49 crc kubenswrapper[4744]: I1201 09:07:49.390922 4744 scope.go:117] "RemoveContainer" containerID="23b9d98ac18b222c3a666b71f499936c00e4bbbeaa0860bc73299c7c08de693f" Dec 01 09:07:49 crc kubenswrapper[4744]: I1201 09:07:49.394066 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:07:49 crc kubenswrapper[4744]: E1201 09:07:49.395802 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:07:49 crc kubenswrapper[4744]: I1201 09:07:49.419860 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=43.270542503 podStartE2EDuration="46.419835093s" podCreationTimestamp="2025-12-01 09:07:03 +0000 UTC" firstStartedPulling="2025-12-01 09:07:03.82356353 +0000 UTC m=+2955.812621451" lastFinishedPulling="2025-12-01 09:07:06.97285612 +0000 UTC m=+2958.961914041" observedRunningTime="2025-12-01 09:07:08.001517501 +0000 UTC m=+2959.990575422" watchObservedRunningTime="2025-12-01 09:07:49.419835093 +0000 UTC m=+3001.408893034" Dec 01 09:08:00 crc kubenswrapper[4744]: I1201 09:08:00.285329 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:08:00 crc kubenswrapper[4744]: E1201 09:08:00.286127 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:08:12 crc kubenswrapper[4744]: I1201 09:08:12.287227 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:08:12 crc kubenswrapper[4744]: E1201 09:08:12.288299 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:08:27 crc kubenswrapper[4744]: I1201 09:08:27.285818 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:08:27 crc kubenswrapper[4744]: E1201 09:08:27.286588 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:08:39 crc kubenswrapper[4744]: I1201 09:08:39.284884 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:08:39 crc kubenswrapper[4744]: E1201 09:08:39.286642 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:08:50 crc kubenswrapper[4744]: I1201 09:08:50.285325 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:08:50 crc kubenswrapper[4744]: E1201 09:08:50.286369 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:09:02 crc kubenswrapper[4744]: I1201 09:09:02.285189 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:09:02 crc kubenswrapper[4744]: E1201 09:09:02.285878 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:09:13 crc kubenswrapper[4744]: I1201 09:09:13.285870 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:09:13 crc kubenswrapper[4744]: E1201 09:09:13.286962 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:09:26 crc kubenswrapper[4744]: I1201 09:09:26.285256 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:09:26 crc kubenswrapper[4744]: E1201 09:09:26.286115 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:09:37 crc kubenswrapper[4744]: I1201 09:09:37.285222 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:09:37 crc kubenswrapper[4744]: E1201 09:09:37.286238 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:09:49 crc kubenswrapper[4744]: I1201 09:09:49.285910 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:09:49 crc kubenswrapper[4744]: E1201 09:09:49.286915 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:10:01 crc kubenswrapper[4744]: I1201 09:10:01.285592 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:10:01 crc kubenswrapper[4744]: E1201 09:10:01.286302 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:10:02 crc kubenswrapper[4744]: I1201 09:10:02.111275 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:10:12 crc kubenswrapper[4744]: I1201 09:10:12.284925 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:10:12 crc kubenswrapper[4744]: E1201 09:10:12.285802 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.272812 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx"] Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.274832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.277761 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.303186 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx"] Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.357389 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.357450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.357512 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmp9h\" (UniqueName: \"kubernetes.io/projected/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-kube-api-access-rmp9h\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.459007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.459425 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.459546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.459723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmp9h\" (UniqueName: \"kubernetes.io/projected/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-kube-api-access-rmp9h\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.459841 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.484475 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmp9h\" (UniqueName: \"kubernetes.io/projected/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-kube-api-access-rmp9h\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:16 crc kubenswrapper[4744]: I1201 09:10:16.601643 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:17 crc kubenswrapper[4744]: I1201 09:10:17.053223 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx"] Dec 01 09:10:17 crc kubenswrapper[4744]: I1201 09:10:17.142778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" event={"ID":"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb","Type":"ContainerStarted","Data":"e6eb1657845b7d5232336232e4b1d9d577a1e5f132aa6bf2141b1a1eecc50e10"} Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.152524 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerID="6a3aa6c2b616a93d4256238e6ae3af4c5dfed8efa4092ef78811ab6e652a9793" exitCode=0 Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.152615 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" event={"ID":"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb","Type":"ContainerDied","Data":"6a3aa6c2b616a93d4256238e6ae3af4c5dfed8efa4092ef78811ab6e652a9793"} Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.611576 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zpr5d"] Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.613472 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.641336 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zpr5d"] Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.705751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-utilities\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.705880 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9k4t\" (UniqueName: \"kubernetes.io/projected/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-kube-api-access-n9k4t\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.705924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-catalog-content\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.807760 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-catalog-content\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.807961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-utilities\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.808327 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-catalog-content\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.808474 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-utilities\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.808635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9k4t\" (UniqueName: \"kubernetes.io/projected/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-kube-api-access-n9k4t\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.830494 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9k4t\" (UniqueName: \"kubernetes.io/projected/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-kube-api-access-n9k4t\") pod \"redhat-operators-zpr5d\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:18 crc kubenswrapper[4744]: I1201 09:10:18.941584 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:19 crc kubenswrapper[4744]: I1201 09:10:19.420806 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zpr5d"] Dec 01 09:10:20 crc kubenswrapper[4744]: I1201 09:10:20.171608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" event={"ID":"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb","Type":"ContainerStarted","Data":"ad321b6be6eadc955a54b6672db7a7246f7667be85b5c2000f25760cc47e1dda"} Dec 01 09:10:20 crc kubenswrapper[4744]: I1201 09:10:20.173127 4744 generic.go:334] "Generic (PLEG): container finished" podID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerID="8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6" exitCode=0 Dec 01 09:10:20 crc kubenswrapper[4744]: I1201 09:10:20.173185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpr5d" event={"ID":"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c","Type":"ContainerDied","Data":"8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6"} Dec 01 09:10:20 crc kubenswrapper[4744]: I1201 09:10:20.173245 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpr5d" event={"ID":"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c","Type":"ContainerStarted","Data":"2a5190d2f1cf48fe4c0cd13dc7f96c420b73f639422be72e29d63fd30b557354"} Dec 01 09:10:21 crc kubenswrapper[4744]: I1201 09:10:21.189379 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerID="ad321b6be6eadc955a54b6672db7a7246f7667be85b5c2000f25760cc47e1dda" exitCode=0 Dec 01 09:10:21 crc kubenswrapper[4744]: I1201 09:10:21.189451 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" event={"ID":"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb","Type":"ContainerDied","Data":"ad321b6be6eadc955a54b6672db7a7246f7667be85b5c2000f25760cc47e1dda"} Dec 01 09:10:22 crc kubenswrapper[4744]: I1201 09:10:22.203164 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpr5d" event={"ID":"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c","Type":"ContainerStarted","Data":"22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018"} Dec 01 09:10:22 crc kubenswrapper[4744]: I1201 09:10:22.206643 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerID="81ac67db72d0c4fe96dff9e6d348fc9a8e7dd9562d26879d55659d049bc6bfaa" exitCode=0 Dec 01 09:10:22 crc kubenswrapper[4744]: I1201 09:10:22.206704 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" event={"ID":"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb","Type":"ContainerDied","Data":"81ac67db72d0c4fe96dff9e6d348fc9a8e7dd9562d26879d55659d049bc6bfaa"} Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.286142 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:10:23 crc kubenswrapper[4744]: E1201 09:10:23.288451 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.599577 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.611932 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-bundle\") pod \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.611992 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-util\") pod \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.612191 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmp9h\" (UniqueName: \"kubernetes.io/projected/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-kube-api-access-rmp9h\") pod \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\" (UID: \"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb\") " Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.622337 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-kube-api-access-rmp9h" (OuterVolumeSpecName: "kube-api-access-rmp9h") pod "e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" (UID: "e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb"). InnerVolumeSpecName "kube-api-access-rmp9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.623500 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-util" (OuterVolumeSpecName: "util") pod "e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" (UID: "e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.623782 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-bundle" (OuterVolumeSpecName: "bundle") pod "e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" (UID: "e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.714083 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.714122 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-util\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:23 crc kubenswrapper[4744]: I1201 09:10:23.714135 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmp9h\" (UniqueName: \"kubernetes.io/projected/e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb-kube-api-access-rmp9h\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:24 crc kubenswrapper[4744]: I1201 09:10:24.226812 4744 generic.go:334] "Generic (PLEG): container finished" podID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerID="22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018" exitCode=0 Dec 01 09:10:24 crc kubenswrapper[4744]: I1201 09:10:24.226898 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpr5d" event={"ID":"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c","Type":"ContainerDied","Data":"22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018"} Dec 01 09:10:24 crc kubenswrapper[4744]: I1201 09:10:24.230740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" event={"ID":"e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb","Type":"ContainerDied","Data":"e6eb1657845b7d5232336232e4b1d9d577a1e5f132aa6bf2141b1a1eecc50e10"} Dec 01 09:10:24 crc kubenswrapper[4744]: I1201 09:10:24.230773 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6eb1657845b7d5232336232e4b1d9d577a1e5f132aa6bf2141b1a1eecc50e10" Dec 01 09:10:24 crc kubenswrapper[4744]: I1201 09:10:24.230801 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx" Dec 01 09:10:25 crc kubenswrapper[4744]: I1201 09:10:25.242878 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpr5d" event={"ID":"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c","Type":"ContainerStarted","Data":"dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483"} Dec 01 09:10:25 crc kubenswrapper[4744]: I1201 09:10:25.266882 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zpr5d" podStartSLOduration=2.734599127 podStartE2EDuration="7.266861949s" podCreationTimestamp="2025-12-01 09:10:18 +0000 UTC" firstStartedPulling="2025-12-01 09:10:20.174773412 +0000 UTC m=+3152.163831343" lastFinishedPulling="2025-12-01 09:10:24.707036234 +0000 UTC m=+3156.696094165" observedRunningTime="2025-12-01 09:10:25.262160525 +0000 UTC m=+3157.251218446" watchObservedRunningTime="2025-12-01 09:10:25.266861949 +0000 UTC m=+3157.255919870" Dec 01 09:10:28 crc kubenswrapper[4744]: I1201 09:10:28.942025 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:28 crc kubenswrapper[4744]: I1201 09:10:28.942732 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:30 crc kubenswrapper[4744]: I1201 09:10:30.009455 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zpr5d" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="registry-server" probeResult="failure" output=< Dec 01 09:10:30 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 09:10:30 crc kubenswrapper[4744]: > Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.788604 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269"] Dec 01 09:10:34 crc kubenswrapper[4744]: E1201 09:10:34.789697 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerName="util" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.789714 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerName="util" Dec 01 09:10:34 crc kubenswrapper[4744]: E1201 09:10:34.789734 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerName="pull" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.789742 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerName="pull" Dec 01 09:10:34 crc kubenswrapper[4744]: E1201 09:10:34.789787 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerName="extract" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.789796 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerName="extract" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.790004 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb" containerName="extract" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.790831 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.792680 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-2sxqs" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.792922 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.794797 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.806962 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269"] Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.907551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndkr7\" (UniqueName: \"kubernetes.io/projected/d7b03d83-6bcb-4e4d-beb9-d4932e4fab91-kube-api-access-ndkr7\") pod \"obo-prometheus-operator-668cf9dfbb-jt269\" (UID: \"d7b03d83-6bcb-4e4d-beb9-d4932e4fab91\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.909148 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj"] Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.910698 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.917585 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-2hvm9" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.920379 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.928875 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp"] Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.930548 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.942222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj"] Dec 01 09:10:34 crc kubenswrapper[4744]: I1201 09:10:34.972620 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp"] Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.011689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndkr7\" (UniqueName: \"kubernetes.io/projected/d7b03d83-6bcb-4e4d-beb9-d4932e4fab91-kube-api-access-ndkr7\") pod \"obo-prometheus-operator-668cf9dfbb-jt269\" (UID: \"d7b03d83-6bcb-4e4d-beb9-d4932e4fab91\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.011795 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50194db7-90e4-4112-83df-f729548b59ca-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj\" (UID: \"50194db7-90e4-4112-83df-f729548b59ca\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.011900 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50194db7-90e4-4112-83df-f729548b59ca-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj\" (UID: \"50194db7-90e4-4112-83df-f729548b59ca\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.032746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndkr7\" (UniqueName: \"kubernetes.io/projected/d7b03d83-6bcb-4e4d-beb9-d4932e4fab91-kube-api-access-ndkr7\") pod \"obo-prometheus-operator-668cf9dfbb-jt269\" (UID: \"d7b03d83-6bcb-4e4d-beb9-d4932e4fab91\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.113593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50194db7-90e4-4112-83df-f729548b59ca-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj\" (UID: \"50194db7-90e4-4112-83df-f729548b59ca\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.113752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/92c3c976-90f4-4d13-9d88-348767b50827-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp\" (UID: \"92c3c976-90f4-4d13-9d88-348767b50827\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.113872 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50194db7-90e4-4112-83df-f729548b59ca-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj\" (UID: \"50194db7-90e4-4112-83df-f729548b59ca\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.113937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/92c3c976-90f4-4d13-9d88-348767b50827-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp\" (UID: \"92c3c976-90f4-4d13-9d88-348767b50827\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.128949 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/50194db7-90e4-4112-83df-f729548b59ca-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj\" (UID: \"50194db7-90e4-4112-83df-f729548b59ca\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.137139 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.144904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/50194db7-90e4-4112-83df-f729548b59ca-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj\" (UID: \"50194db7-90e4-4112-83df-f729548b59ca\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.180882 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-m2tkm"] Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.182368 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.185944 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-7d88m" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.186320 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.220017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/92c3c976-90f4-4d13-9d88-348767b50827-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp\" (UID: \"92c3c976-90f4-4d13-9d88-348767b50827\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.220208 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/92c3c976-90f4-4d13-9d88-348767b50827-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp\" (UID: \"92c3c976-90f4-4d13-9d88-348767b50827\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.236387 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/92c3c976-90f4-4d13-9d88-348767b50827-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp\" (UID: \"92c3c976-90f4-4d13-9d88-348767b50827\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.237121 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.240495 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-m2tkm"] Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.245955 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/92c3c976-90f4-4d13-9d88-348767b50827-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp\" (UID: \"92c3c976-90f4-4d13-9d88-348767b50827\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.263866 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.323512 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1466d56b-9044-4bd8-97f3-02c1feafa96f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-m2tkm\" (UID: \"1466d56b-9044-4bd8-97f3-02c1feafa96f\") " pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.323703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f55wn\" (UniqueName: \"kubernetes.io/projected/1466d56b-9044-4bd8-97f3-02c1feafa96f-kube-api-access-f55wn\") pod \"observability-operator-d8bb48f5d-m2tkm\" (UID: \"1466d56b-9044-4bd8-97f3-02c1feafa96f\") " pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.402698 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tk4rw"] Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.414599 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.422386 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-rgpgn" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.435076 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1466d56b-9044-4bd8-97f3-02c1feafa96f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-m2tkm\" (UID: \"1466d56b-9044-4bd8-97f3-02c1feafa96f\") " pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.435445 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f55wn\" (UniqueName: \"kubernetes.io/projected/1466d56b-9044-4bd8-97f3-02c1feafa96f-kube-api-access-f55wn\") pod \"observability-operator-d8bb48f5d-m2tkm\" (UID: \"1466d56b-9044-4bd8-97f3-02c1feafa96f\") " pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.443198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1466d56b-9044-4bd8-97f3-02c1feafa96f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-m2tkm\" (UID: \"1466d56b-9044-4bd8-97f3-02c1feafa96f\") " pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.467302 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tk4rw"] Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.481601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f55wn\" (UniqueName: \"kubernetes.io/projected/1466d56b-9044-4bd8-97f3-02c1feafa96f-kube-api-access-f55wn\") pod \"observability-operator-d8bb48f5d-m2tkm\" (UID: \"1466d56b-9044-4bd8-97f3-02c1feafa96f\") " pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.540182 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d48e874b-8560-49fa-9981-136edeec81f7-openshift-service-ca\") pod \"perses-operator-5446b9c989-tk4rw\" (UID: \"d48e874b-8560-49fa-9981-136edeec81f7\") " pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.540218 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxkjm\" (UniqueName: \"kubernetes.io/projected/d48e874b-8560-49fa-9981-136edeec81f7-kube-api-access-sxkjm\") pod \"perses-operator-5446b9c989-tk4rw\" (UID: \"d48e874b-8560-49fa-9981-136edeec81f7\") " pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.633317 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.641540 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d48e874b-8560-49fa-9981-136edeec81f7-openshift-service-ca\") pod \"perses-operator-5446b9c989-tk4rw\" (UID: \"d48e874b-8560-49fa-9981-136edeec81f7\") " pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.641587 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxkjm\" (UniqueName: \"kubernetes.io/projected/d48e874b-8560-49fa-9981-136edeec81f7-kube-api-access-sxkjm\") pod \"perses-operator-5446b9c989-tk4rw\" (UID: \"d48e874b-8560-49fa-9981-136edeec81f7\") " pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.642335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d48e874b-8560-49fa-9981-136edeec81f7-openshift-service-ca\") pod \"perses-operator-5446b9c989-tk4rw\" (UID: \"d48e874b-8560-49fa-9981-136edeec81f7\") " pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.682815 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxkjm\" (UniqueName: \"kubernetes.io/projected/d48e874b-8560-49fa-9981-136edeec81f7-kube-api-access-sxkjm\") pod \"perses-operator-5446b9c989-tk4rw\" (UID: \"d48e874b-8560-49fa-9981-136edeec81f7\") " pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:35 crc kubenswrapper[4744]: I1201 09:10:35.765672 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.125940 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj"] Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.144156 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269"] Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.285535 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:10:36 crc kubenswrapper[4744]: E1201 09:10:36.285817 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.319987 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp"] Dec 01 09:10:36 crc kubenswrapper[4744]: W1201 09:10:36.346675 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92c3c976_90f4_4d13_9d88_348767b50827.slice/crio-028d998829271b52953786629f42329a520f2506879438331ab4785421646c23 WatchSource:0}: Error finding container 028d998829271b52953786629f42329a520f2506879438331ab4785421646c23: Status 404 returned error can't find the container with id 028d998829271b52953786629f42329a520f2506879438331ab4785421646c23 Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.355194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" event={"ID":"d7b03d83-6bcb-4e4d-beb9-d4932e4fab91","Type":"ContainerStarted","Data":"25af8a65ad781082f0a7e359da03b1cd01438a7aa9abc41b68cc89e36f5a7d24"} Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.360385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" event={"ID":"50194db7-90e4-4112-83df-f729548b59ca","Type":"ContainerStarted","Data":"fadbe7bf2050b78fe87d7caadc651e0a71110f504b2006bf4b8a8f478f4bc046"} Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.500884 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tk4rw"] Dec 01 09:10:36 crc kubenswrapper[4744]: I1201 09:10:36.511215 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-m2tkm"] Dec 01 09:10:37 crc kubenswrapper[4744]: I1201 09:10:37.377108 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" event={"ID":"92c3c976-90f4-4d13-9d88-348767b50827","Type":"ContainerStarted","Data":"028d998829271b52953786629f42329a520f2506879438331ab4785421646c23"} Dec 01 09:10:37 crc kubenswrapper[4744]: I1201 09:10:37.378979 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" event={"ID":"1466d56b-9044-4bd8-97f3-02c1feafa96f","Type":"ContainerStarted","Data":"740f7bd544f0da3653ea72bf47bbaac6e74fb34a5abb2dde84db1ba85b067551"} Dec 01 09:10:37 crc kubenswrapper[4744]: I1201 09:10:37.380565 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-tk4rw" event={"ID":"d48e874b-8560-49fa-9981-136edeec81f7","Type":"ContainerStarted","Data":"e5503d17699f4a3964b7aa3eab59e0d830c5eaf3994eb15f41418122058435ad"} Dec 01 09:10:40 crc kubenswrapper[4744]: I1201 09:10:40.007497 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zpr5d" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="registry-server" probeResult="failure" output=< Dec 01 09:10:40 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 09:10:40 crc kubenswrapper[4744]: > Dec 01 09:10:49 crc kubenswrapper[4744]: I1201 09:10:49.001724 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:49 crc kubenswrapper[4744]: I1201 09:10:49.140624 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:50 crc kubenswrapper[4744]: I1201 09:10:50.285587 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:10:50 crc kubenswrapper[4744]: E1201 09:10:50.286314 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:10:51 crc kubenswrapper[4744]: I1201 09:10:51.002399 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zpr5d"] Dec 01 09:10:51 crc kubenswrapper[4744]: I1201 09:10:51.002671 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zpr5d" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="registry-server" containerID="cri-o://dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483" gracePeriod=2 Dec 01 09:10:51 crc kubenswrapper[4744]: E1201 09:10:51.791228 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 01 09:10:51 crc kubenswrapper[4744]: E1201 09:10:51.791748 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp_openshift-operators(92c3c976-90f4-4d13-9d88-348767b50827): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 09:10:51 crc kubenswrapper[4744]: E1201 09:10:51.793635 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" podUID="92c3c976-90f4-4d13-9d88-348767b50827" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.233977 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.323797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-catalog-content\") pod \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.323946 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-utilities\") pod \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.324021 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9k4t\" (UniqueName: \"kubernetes.io/projected/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-kube-api-access-n9k4t\") pod \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\" (UID: \"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c\") " Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.325182 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-utilities" (OuterVolumeSpecName: "utilities") pod "d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" (UID: "d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.377815 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-kube-api-access-n9k4t" (OuterVolumeSpecName: "kube-api-access-n9k4t") pod "d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" (UID: "d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c"). InnerVolumeSpecName "kube-api-access-n9k4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.427269 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.427300 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9k4t\" (UniqueName: \"kubernetes.io/projected/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-kube-api-access-n9k4t\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.453443 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" (UID: "d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.528937 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.610839 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" event={"ID":"50194db7-90e4-4112-83df-f729548b59ca","Type":"ContainerStarted","Data":"c64b27269f51fb74479a4f63fc437826f622ebcb9297476c7ed9ae54f3a6e1c2"} Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.612689 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" event={"ID":"1466d56b-9044-4bd8-97f3-02c1feafa96f","Type":"ContainerStarted","Data":"790a05f79ffd9c651de11e9c0696b011bcbe0f97c60fd84b37b40acf0fb51d7f"} Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.615190 4744 generic.go:334] "Generic (PLEG): container finished" podID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerID="dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483" exitCode=0 Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.615262 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpr5d" event={"ID":"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c","Type":"ContainerDied","Data":"dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483"} Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.615336 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zpr5d" event={"ID":"d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c","Type":"ContainerDied","Data":"2a5190d2f1cf48fe4c0cd13dc7f96c420b73f639422be72e29d63fd30b557354"} Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.615362 4744 scope.go:117] "RemoveContainer" containerID="dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.615541 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zpr5d" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.617023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" event={"ID":"d7b03d83-6bcb-4e4d-beb9-d4932e4fab91","Type":"ContainerStarted","Data":"db21e6c9e89dd4edc9948f3422e7770c3b5d58b656fbbbfbcfc63c55d65e756d"} Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.618960 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-tk4rw" event={"ID":"d48e874b-8560-49fa-9981-136edeec81f7","Type":"ContainerStarted","Data":"5a3fed97f734c947c078b97bdb6cc4e1e11a4fac254485a896f23d46bba9f676"} Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.619110 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.649632 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj" podStartSLOduration=2.967231238 podStartE2EDuration="18.649600499s" podCreationTimestamp="2025-12-01 09:10:34 +0000 UTC" firstStartedPulling="2025-12-01 09:10:36.124202173 +0000 UTC m=+3168.113260094" lastFinishedPulling="2025-12-01 09:10:51.806571434 +0000 UTC m=+3183.795629355" observedRunningTime="2025-12-01 09:10:52.640209542 +0000 UTC m=+3184.629267473" watchObservedRunningTime="2025-12-01 09:10:52.649600499 +0000 UTC m=+3184.638658420" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.649723 4744 scope.go:117] "RemoveContainer" containerID="22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.727093 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zpr5d"] Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.759752 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zpr5d"] Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.776631 4744 scope.go:117] "RemoveContainer" containerID="8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.821540 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" podStartSLOduration=2.445740399 podStartE2EDuration="17.821514443s" podCreationTimestamp="2025-12-01 09:10:35 +0000 UTC" firstStartedPulling="2025-12-01 09:10:36.498387774 +0000 UTC m=+3168.487445695" lastFinishedPulling="2025-12-01 09:10:51.874161818 +0000 UTC m=+3183.863219739" observedRunningTime="2025-12-01 09:10:52.790533841 +0000 UTC m=+3184.779591762" watchObservedRunningTime="2025-12-01 09:10:52.821514443 +0000 UTC m=+3184.810572364" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.877370 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jt269" podStartSLOduration=3.222293499 podStartE2EDuration="18.877349692s" podCreationTimestamp="2025-12-01 09:10:34 +0000 UTC" firstStartedPulling="2025-12-01 09:10:36.139689274 +0000 UTC m=+3168.128747195" lastFinishedPulling="2025-12-01 09:10:51.794745457 +0000 UTC m=+3183.783803388" observedRunningTime="2025-12-01 09:10:52.830748515 +0000 UTC m=+3184.819806446" watchObservedRunningTime="2025-12-01 09:10:52.877349692 +0000 UTC m=+3184.866407613" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.880184 4744 scope.go:117] "RemoveContainer" containerID="dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483" Dec 01 09:10:52 crc kubenswrapper[4744]: E1201 09:10:52.880597 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483\": container with ID starting with dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483 not found: ID does not exist" containerID="dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.880628 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483"} err="failed to get container status \"dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483\": rpc error: code = NotFound desc = could not find container \"dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483\": container with ID starting with dedec32b4fa36cf124f38e9025c49f6c62062d0a09d244f6b075112bc7649483 not found: ID does not exist" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.880653 4744 scope.go:117] "RemoveContainer" containerID="22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018" Dec 01 09:10:52 crc kubenswrapper[4744]: E1201 09:10:52.880880 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018\": container with ID starting with 22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018 not found: ID does not exist" containerID="22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.880902 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018"} err="failed to get container status \"22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018\": rpc error: code = NotFound desc = could not find container \"22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018\": container with ID starting with 22ac1f38cd695249faf1b06ba785726ab13e623f9b5a72590c979cc9669bc018 not found: ID does not exist" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.880919 4744 scope.go:117] "RemoveContainer" containerID="8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6" Dec 01 09:10:52 crc kubenswrapper[4744]: E1201 09:10:52.881119 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6\": container with ID starting with 8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6 not found: ID does not exist" containerID="8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.881155 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6"} err="failed to get container status \"8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6\": rpc error: code = NotFound desc = could not find container \"8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6\": container with ID starting with 8dd27c87ba3c64c365057cd54c3f303f96f61cf7787469544fed2a90fc2eedd6 not found: ID does not exist" Dec 01 09:10:52 crc kubenswrapper[4744]: I1201 09:10:52.881474 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-tk4rw" podStartSLOduration=2.567357521 podStartE2EDuration="17.881455789s" podCreationTimestamp="2025-12-01 09:10:35 +0000 UTC" firstStartedPulling="2025-12-01 09:10:36.493197886 +0000 UTC m=+3168.482255807" lastFinishedPulling="2025-12-01 09:10:51.807296144 +0000 UTC m=+3183.796354075" observedRunningTime="2025-12-01 09:10:52.862939832 +0000 UTC m=+3184.851997773" watchObservedRunningTime="2025-12-01 09:10:52.881455789 +0000 UTC m=+3184.870513700" Dec 01 09:10:53 crc kubenswrapper[4744]: I1201 09:10:53.646748 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" event={"ID":"92c3c976-90f4-4d13-9d88-348767b50827","Type":"ContainerStarted","Data":"e7efe49a5e389a66d18d2af614e05168fd72b23e59eb728319cd68cc2a528042"} Dec 01 09:10:53 crc kubenswrapper[4744]: I1201 09:10:53.648648 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:53 crc kubenswrapper[4744]: I1201 09:10:53.652901 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-m2tkm" Dec 01 09:10:53 crc kubenswrapper[4744]: I1201 09:10:53.672092 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp" podStartSLOduration=-9223372017.1827 podStartE2EDuration="19.672076022s" podCreationTimestamp="2025-12-01 09:10:34 +0000 UTC" firstStartedPulling="2025-12-01 09:10:36.350768722 +0000 UTC m=+3168.339826643" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:10:53.665395971 +0000 UTC m=+3185.654453892" watchObservedRunningTime="2025-12-01 09:10:53.672076022 +0000 UTC m=+3185.661133943" Dec 01 09:10:54 crc kubenswrapper[4744]: I1201 09:10:54.295465 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" path="/var/lib/kubelet/pods/d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c/volumes" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.030848 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hrhlm"] Dec 01 09:10:57 crc kubenswrapper[4744]: E1201 09:10:57.031845 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="registry-server" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.031861 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="registry-server" Dec 01 09:10:57 crc kubenswrapper[4744]: E1201 09:10:57.031878 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="extract-utilities" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.031886 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="extract-utilities" Dec 01 09:10:57 crc kubenswrapper[4744]: E1201 09:10:57.031917 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="extract-content" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.031924 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="extract-content" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.032169 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b0fe58-3ff0-4efa-a1ca-d7f6b18a542c" containerName="registry-server" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.038220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.049242 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrhlm"] Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.155902 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-catalog-content\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.155965 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-utilities\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.155986 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmnkd\" (UniqueName: \"kubernetes.io/projected/17d20c0c-e648-4e45-bab4-22a808ff3089-kube-api-access-fmnkd\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.257188 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-catalog-content\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.257251 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-utilities\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.257275 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmnkd\" (UniqueName: \"kubernetes.io/projected/17d20c0c-e648-4e45-bab4-22a808ff3089-kube-api-access-fmnkd\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.257826 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-catalog-content\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.257991 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-utilities\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.284117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmnkd\" (UniqueName: \"kubernetes.io/projected/17d20c0c-e648-4e45-bab4-22a808ff3089-kube-api-access-fmnkd\") pod \"certified-operators-hrhlm\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.359134 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:10:57 crc kubenswrapper[4744]: I1201 09:10:57.916542 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrhlm"] Dec 01 09:10:58 crc kubenswrapper[4744]: I1201 09:10:58.699904 4744 generic.go:334] "Generic (PLEG): container finished" podID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerID="f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc" exitCode=0 Dec 01 09:10:58 crc kubenswrapper[4744]: I1201 09:10:58.700497 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhlm" event={"ID":"17d20c0c-e648-4e45-bab4-22a808ff3089","Type":"ContainerDied","Data":"f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc"} Dec 01 09:10:58 crc kubenswrapper[4744]: I1201 09:10:58.700523 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhlm" event={"ID":"17d20c0c-e648-4e45-bab4-22a808ff3089","Type":"ContainerStarted","Data":"90ea3e8e4fb6e8fb424cac7ec7f9ef09449817468f7599b720a32f54aacc9d81"} Dec 01 09:11:00 crc kubenswrapper[4744]: I1201 09:11:00.721258 4744 generic.go:334] "Generic (PLEG): container finished" podID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerID="ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb" exitCode=0 Dec 01 09:11:00 crc kubenswrapper[4744]: I1201 09:11:00.721313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhlm" event={"ID":"17d20c0c-e648-4e45-bab4-22a808ff3089","Type":"ContainerDied","Data":"ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb"} Dec 01 09:11:00 crc kubenswrapper[4744]: I1201 09:11:00.966537 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:11:00 crc kubenswrapper[4744]: I1201 09:11:00.967039 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-api" containerID="cri-o://3073308ea771cc8aa2f12ffa6d73e302ce7684c382d0bf167d592bb0b9ff25ee" gracePeriod=30 Dec 01 09:11:00 crc kubenswrapper[4744]: I1201 09:11:00.967133 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-listener" containerID="cri-o://a19b8319fa651e231d9ad0039d9b673631d193dec59c60c3c894d33217456a0c" gracePeriod=30 Dec 01 09:11:00 crc kubenswrapper[4744]: I1201 09:11:00.967183 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-notifier" containerID="cri-o://776d6f0bac2b8d73de7f32ce04a1bec2e5dc5d95d742941e6d3a5b548818c1b0" gracePeriod=30 Dec 01 09:11:00 crc kubenswrapper[4744]: I1201 09:11:00.967195 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-evaluator" containerID="cri-o://8a8dd1d6f9777ebaa902f10e5b908ba2c65e7651dfcc963db3021e84fe0c725a" gracePeriod=30 Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.536078 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.538242 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.540123 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.541032 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-tcdkq" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.541239 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.541440 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.541793 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.561021 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.698555 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zwkr\" (UniqueName: \"kubernetes.io/projected/0c022f75-f489-4eee-a701-f19067fe78a2-kube-api-access-9zwkr\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.699206 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.699487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0c022f75-f489-4eee-a701-f19067fe78a2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.699584 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.699756 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.700058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c022f75-f489-4eee-a701-f19067fe78a2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.700319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c022f75-f489-4eee-a701-f19067fe78a2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.743240 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhlm" event={"ID":"17d20c0c-e648-4e45-bab4-22a808ff3089","Type":"ContainerStarted","Data":"aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457"} Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.750241 4744 generic.go:334] "Generic (PLEG): container finished" podID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerID="8a8dd1d6f9777ebaa902f10e5b908ba2c65e7651dfcc963db3021e84fe0c725a" exitCode=0 Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.750283 4744 generic.go:334] "Generic (PLEG): container finished" podID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerID="3073308ea771cc8aa2f12ffa6d73e302ce7684c382d0bf167d592bb0b9ff25ee" exitCode=0 Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.750300 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerDied","Data":"8a8dd1d6f9777ebaa902f10e5b908ba2c65e7651dfcc963db3021e84fe0c725a"} Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.750378 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerDied","Data":"3073308ea771cc8aa2f12ffa6d73e302ce7684c382d0bf167d592bb0b9ff25ee"} Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.772978 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hrhlm" podStartSLOduration=3.007887876 podStartE2EDuration="5.772955049s" podCreationTimestamp="2025-12-01 09:10:56 +0000 UTC" firstStartedPulling="2025-12-01 09:10:58.702745461 +0000 UTC m=+3190.691803382" lastFinishedPulling="2025-12-01 09:11:01.467812634 +0000 UTC m=+3193.456870555" observedRunningTime="2025-12-01 09:11:01.761570505 +0000 UTC m=+3193.750628436" watchObservedRunningTime="2025-12-01 09:11:01.772955049 +0000 UTC m=+3193.762012970" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.801925 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c022f75-f489-4eee-a701-f19067fe78a2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.802016 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c022f75-f489-4eee-a701-f19067fe78a2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.802055 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zwkr\" (UniqueName: \"kubernetes.io/projected/0c022f75-f489-4eee-a701-f19067fe78a2-kube-api-access-9zwkr\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.802099 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.802139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0c022f75-f489-4eee-a701-f19067fe78a2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.802159 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.802193 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.802890 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/0c022f75-f489-4eee-a701-f19067fe78a2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.812219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.812219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.812569 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c022f75-f489-4eee-a701-f19067fe78a2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.813032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0c022f75-f489-4eee-a701-f19067fe78a2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.815744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c022f75-f489-4eee-a701-f19067fe78a2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.837643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zwkr\" (UniqueName: \"kubernetes.io/projected/0c022f75-f489-4eee-a701-f19067fe78a2-kube-api-access-9zwkr\") pod \"alertmanager-metric-storage-0\" (UID: \"0c022f75-f489-4eee-a701-f19067fe78a2\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:01 crc kubenswrapper[4744]: I1201 09:11:01.976029 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.174567 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.177372 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.184466 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.185976 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fdbml" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.186214 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.186362 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.185974 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.186644 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.186766 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.288639 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:11:02 crc kubenswrapper[4744]: E1201 09:11:02.288960 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.312989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.313069 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88k8g\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-kube-api-access-88k8g\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.313112 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.313151 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.313205 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.313236 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f40458d8-65cc-49fe-bbcb-16f0b4290d70-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.313302 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.313338 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.416170 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.416750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88k8g\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-kube-api-access-88k8g\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.418362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.418496 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.419105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.419186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f40458d8-65cc-49fe-bbcb-16f0b4290d70-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.419762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.420246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.421070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f40458d8-65cc-49fe-bbcb-16f0b4290d70-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.428633 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.432240 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.440814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.442066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.442582 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.445434 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.448314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88k8g\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-kube-api-access-88k8g\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.489349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:02 crc kubenswrapper[4744]: W1201 09:11:02.542980 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c022f75_f489_4eee_a701_f19067fe78a2.slice/crio-553e206525ffed029cf228d14b41acd1024319fb44fa21c2d1cba255e4cfa120 WatchSource:0}: Error finding container 553e206525ffed029cf228d14b41acd1024319fb44fa21c2d1cba255e4cfa120: Status 404 returned error can't find the container with id 553e206525ffed029cf228d14b41acd1024319fb44fa21c2d1cba255e4cfa120 Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.544502 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 09:11:02 crc kubenswrapper[4744]: I1201 09:11:02.557250 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:03 crc kubenswrapper[4744]: I1201 09:11:02.810702 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0c022f75-f489-4eee-a701-f19067fe78a2","Type":"ContainerStarted","Data":"553e206525ffed029cf228d14b41acd1024319fb44fa21c2d1cba255e4cfa120"} Dec 01 09:11:03 crc kubenswrapper[4744]: I1201 09:11:03.784664 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:03 crc kubenswrapper[4744]: I1201 09:11:03.831166 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerStarted","Data":"a5ced283d875c2fab6b59793fafeca56136afb9cf63400554096a76cd1488303"} Dec 01 09:11:04 crc kubenswrapper[4744]: I1201 09:11:04.845808 4744 generic.go:334] "Generic (PLEG): container finished" podID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerID="a19b8319fa651e231d9ad0039d9b673631d193dec59c60c3c894d33217456a0c" exitCode=0 Dec 01 09:11:04 crc kubenswrapper[4744]: I1201 09:11:04.846082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerDied","Data":"a19b8319fa651e231d9ad0039d9b673631d193dec59c60c3c894d33217456a0c"} Dec 01 09:11:05 crc kubenswrapper[4744]: I1201 09:11:05.768567 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-tk4rw" Dec 01 09:11:07 crc kubenswrapper[4744]: I1201 09:11:07.359861 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:11:07 crc kubenswrapper[4744]: I1201 09:11:07.360172 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:11:07 crc kubenswrapper[4744]: I1201 09:11:07.416915 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:11:07 crc kubenswrapper[4744]: I1201 09:11:07.919893 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:11:08 crc kubenswrapper[4744]: I1201 09:11:08.882257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerStarted","Data":"88ebf2a52c1ee892f132dcfdefc4dcdf0b44c1ac03e633702961be7a0fdbfd86"} Dec 01 09:11:08 crc kubenswrapper[4744]: I1201 09:11:08.883716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0c022f75-f489-4eee-a701-f19067fe78a2","Type":"ContainerStarted","Data":"28d9e1a124510ba72a60081519184bd8d80abba7427894de7c772867c57b5e3a"} Dec 01 09:11:09 crc kubenswrapper[4744]: I1201 09:11:09.802129 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrhlm"] Dec 01 09:11:09 crc kubenswrapper[4744]: I1201 09:11:09.892537 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hrhlm" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="registry-server" containerID="cri-o://aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457" gracePeriod=2 Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.576940 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.711066 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-catalog-content\") pod \"17d20c0c-e648-4e45-bab4-22a808ff3089\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.711200 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-utilities\") pod \"17d20c0c-e648-4e45-bab4-22a808ff3089\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.711319 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmnkd\" (UniqueName: \"kubernetes.io/projected/17d20c0c-e648-4e45-bab4-22a808ff3089-kube-api-access-fmnkd\") pod \"17d20c0c-e648-4e45-bab4-22a808ff3089\" (UID: \"17d20c0c-e648-4e45-bab4-22a808ff3089\") " Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.712478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-utilities" (OuterVolumeSpecName: "utilities") pod "17d20c0c-e648-4e45-bab4-22a808ff3089" (UID: "17d20c0c-e648-4e45-bab4-22a808ff3089"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.722695 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d20c0c-e648-4e45-bab4-22a808ff3089-kube-api-access-fmnkd" (OuterVolumeSpecName: "kube-api-access-fmnkd") pod "17d20c0c-e648-4e45-bab4-22a808ff3089" (UID: "17d20c0c-e648-4e45-bab4-22a808ff3089"). InnerVolumeSpecName "kube-api-access-fmnkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.813368 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmnkd\" (UniqueName: \"kubernetes.io/projected/17d20c0c-e648-4e45-bab4-22a808ff3089-kube-api-access-fmnkd\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.813399 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.902293 4744 generic.go:334] "Generic (PLEG): container finished" podID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerID="aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457" exitCode=0 Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.902338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhlm" event={"ID":"17d20c0c-e648-4e45-bab4-22a808ff3089","Type":"ContainerDied","Data":"aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457"} Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.902366 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrhlm" event={"ID":"17d20c0c-e648-4e45-bab4-22a808ff3089","Type":"ContainerDied","Data":"90ea3e8e4fb6e8fb424cac7ec7f9ef09449817468f7599b720a32f54aacc9d81"} Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.902386 4744 scope.go:117] "RemoveContainer" containerID="aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.902561 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrhlm" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.927957 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17d20c0c-e648-4e45-bab4-22a808ff3089" (UID: "17d20c0c-e648-4e45-bab4-22a808ff3089"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.938507 4744 scope.go:117] "RemoveContainer" containerID="ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb" Dec 01 09:11:10 crc kubenswrapper[4744]: I1201 09:11:10.963885 4744 scope.go:117] "RemoveContainer" containerID="f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.019031 4744 scope.go:117] "RemoveContainer" containerID="aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.020242 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d20c0c-e648-4e45-bab4-22a808ff3089-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:11 crc kubenswrapper[4744]: E1201 09:11:11.021174 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457\": container with ID starting with aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457 not found: ID does not exist" containerID="aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.021226 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457"} err="failed to get container status \"aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457\": rpc error: code = NotFound desc = could not find container \"aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457\": container with ID starting with aac79fbdbeb8135064d1a532164e7a7c408bfbbc11e40d99d1543675e0008457 not found: ID does not exist" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.021258 4744 scope.go:117] "RemoveContainer" containerID="ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb" Dec 01 09:11:11 crc kubenswrapper[4744]: E1201 09:11:11.021599 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb\": container with ID starting with ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb not found: ID does not exist" containerID="ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.021624 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb"} err="failed to get container status \"ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb\": rpc error: code = NotFound desc = could not find container \"ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb\": container with ID starting with ab68ae772def1b1bc929636f8d7b969fda84cf4dc7750b2ab851ad8881d4acfb not found: ID does not exist" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.021639 4744 scope.go:117] "RemoveContainer" containerID="f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc" Dec 01 09:11:11 crc kubenswrapper[4744]: E1201 09:11:11.021952 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc\": container with ID starting with f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc not found: ID does not exist" containerID="f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.021978 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc"} err="failed to get container status \"f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc\": rpc error: code = NotFound desc = could not find container \"f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc\": container with ID starting with f5e59165822e1c7c416d33ec3be85b76a0836c04a0c04b206dcd61c3e968ebfc not found: ID does not exist" Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.252238 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrhlm"] Dec 01 09:11:11 crc kubenswrapper[4744]: I1201 09:11:11.263841 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hrhlm"] Dec 01 09:11:12 crc kubenswrapper[4744]: I1201 09:11:12.298909 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" path="/var/lib/kubelet/pods/17d20c0c-e648-4e45-bab4-22a808ff3089/volumes" Dec 01 09:11:14 crc kubenswrapper[4744]: I1201 09:11:14.941038 4744 generic.go:334] "Generic (PLEG): container finished" podID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerID="88ebf2a52c1ee892f132dcfdefc4dcdf0b44c1ac03e633702961be7a0fdbfd86" exitCode=0 Dec 01 09:11:14 crc kubenswrapper[4744]: I1201 09:11:14.941118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerDied","Data":"88ebf2a52c1ee892f132dcfdefc4dcdf0b44c1ac03e633702961be7a0fdbfd86"} Dec 01 09:11:14 crc kubenswrapper[4744]: I1201 09:11:14.943596 4744 generic.go:334] "Generic (PLEG): container finished" podID="0c022f75-f489-4eee-a701-f19067fe78a2" containerID="28d9e1a124510ba72a60081519184bd8d80abba7427894de7c772867c57b5e3a" exitCode=0 Dec 01 09:11:14 crc kubenswrapper[4744]: I1201 09:11:14.943643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0c022f75-f489-4eee-a701-f19067fe78a2","Type":"ContainerDied","Data":"28d9e1a124510ba72a60081519184bd8d80abba7427894de7c772867c57b5e3a"} Dec 01 09:11:14 crc kubenswrapper[4744]: I1201 09:11:14.949855 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:11:15 crc kubenswrapper[4744]: I1201 09:11:15.284859 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:11:15 crc kubenswrapper[4744]: E1201 09:11:15.285511 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:11:17 crc kubenswrapper[4744]: I1201 09:11:17.991693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0c022f75-f489-4eee-a701-f19067fe78a2","Type":"ContainerStarted","Data":"22803c75d9a330b44510bbb7482f51570794e3c0df23b50de7561d592f5f73b8"} Dec 01 09:11:21 crc kubenswrapper[4744]: I1201 09:11:21.023800 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"0c022f75-f489-4eee-a701-f19067fe78a2","Type":"ContainerStarted","Data":"aba12291cdc8f4f1dac3289e646ef75ce01c324d426ab2d243e951400bec3403"} Dec 01 09:11:21 crc kubenswrapper[4744]: I1201 09:11:21.025654 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:21 crc kubenswrapper[4744]: I1201 09:11:21.028636 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 01 09:11:21 crc kubenswrapper[4744]: I1201 09:11:21.047486 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.505266436 podStartE2EDuration="20.047470393s" podCreationTimestamp="2025-12-01 09:11:01 +0000 UTC" firstStartedPulling="2025-12-01 09:11:02.545299893 +0000 UTC m=+3194.534357814" lastFinishedPulling="2025-12-01 09:11:17.08750385 +0000 UTC m=+3209.076561771" observedRunningTime="2025-12-01 09:11:21.044013945 +0000 UTC m=+3213.033071876" watchObservedRunningTime="2025-12-01 09:11:21.047470393 +0000 UTC m=+3213.036528314" Dec 01 09:11:23 crc kubenswrapper[4744]: I1201 09:11:23.048134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerStarted","Data":"2dd49f491993313e4adc9252a560e841527776d85de0248ec2415b1e213e27eb"} Dec 01 09:11:26 crc kubenswrapper[4744]: I1201 09:11:26.084331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerStarted","Data":"c2997a70a5b15362f5b693221990bd55322d1ac5e2ca0352d4f1905094c72211"} Dec 01 09:11:28 crc kubenswrapper[4744]: I1201 09:11:28.291824 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:11:28 crc kubenswrapper[4744]: E1201 09:11:28.293319 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.154820 4744 generic.go:334] "Generic (PLEG): container finished" podID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerID="776d6f0bac2b8d73de7f32ce04a1bec2e5dc5d95d742941e6d3a5b548818c1b0" exitCode=137 Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.154871 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerDied","Data":"776d6f0bac2b8d73de7f32ce04a1bec2e5dc5d95d742941e6d3a5b548818c1b0"} Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.159834 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerStarted","Data":"e4906d1a18342d363d9e8f79ec957a8c90ccebb1d1fe52fed2b03d4657ca030c"} Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.191311 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.629106574 podStartE2EDuration="30.191293489s" podCreationTimestamp="2025-12-01 09:11:01 +0000 UTC" firstStartedPulling="2025-12-01 09:11:03.791289587 +0000 UTC m=+3195.780347548" lastFinishedPulling="2025-12-01 09:11:30.353476542 +0000 UTC m=+3222.342534463" observedRunningTime="2025-12-01 09:11:31.185235427 +0000 UTC m=+3223.174293358" watchObservedRunningTime="2025-12-01 09:11:31.191293489 +0000 UTC m=+3223.180351410" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.520856 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.563166 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-scripts\") pod \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.563300 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-internal-tls-certs\") pod \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.563510 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-combined-ca-bundle\") pod \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.563621 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qjpt\" (UniqueName: \"kubernetes.io/projected/ebb6d047-293a-42e1-9dd2-c35a736f93ea-kube-api-access-4qjpt\") pod \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.563665 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-public-tls-certs\") pod \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.563728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-config-data\") pod \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\" (UID: \"ebb6d047-293a-42e1-9dd2-c35a736f93ea\") " Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.571563 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebb6d047-293a-42e1-9dd2-c35a736f93ea-kube-api-access-4qjpt" (OuterVolumeSpecName: "kube-api-access-4qjpt") pod "ebb6d047-293a-42e1-9dd2-c35a736f93ea" (UID: "ebb6d047-293a-42e1-9dd2-c35a736f93ea"). InnerVolumeSpecName "kube-api-access-4qjpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.572599 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-scripts" (OuterVolumeSpecName: "scripts") pod "ebb6d047-293a-42e1-9dd2-c35a736f93ea" (UID: "ebb6d047-293a-42e1-9dd2-c35a736f93ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.648322 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ebb6d047-293a-42e1-9dd2-c35a736f93ea" (UID: "ebb6d047-293a-42e1-9dd2-c35a736f93ea"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.659438 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ebb6d047-293a-42e1-9dd2-c35a736f93ea" (UID: "ebb6d047-293a-42e1-9dd2-c35a736f93ea"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.667795 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qjpt\" (UniqueName: \"kubernetes.io/projected/ebb6d047-293a-42e1-9dd2-c35a736f93ea-kube-api-access-4qjpt\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.667823 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.667835 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.667844 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.691791 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-config-data" (OuterVolumeSpecName: "config-data") pod "ebb6d047-293a-42e1-9dd2-c35a736f93ea" (UID: "ebb6d047-293a-42e1-9dd2-c35a736f93ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.710511 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebb6d047-293a-42e1-9dd2-c35a736f93ea" (UID: "ebb6d047-293a-42e1-9dd2-c35a736f93ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.770691 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:31 crc kubenswrapper[4744]: I1201 09:11:31.770776 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebb6d047-293a-42e1-9dd2-c35a736f93ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.174473 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ebb6d047-293a-42e1-9dd2-c35a736f93ea","Type":"ContainerDied","Data":"225beba64fdafff926b32c013efeae9d35c7fae17ae501f006fa45ef50625cbe"} Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.174802 4744 scope.go:117] "RemoveContainer" containerID="a19b8319fa651e231d9ad0039d9b673631d193dec59c60c3c894d33217456a0c" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.174547 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.220311 4744 scope.go:117] "RemoveContainer" containerID="776d6f0bac2b8d73de7f32ce04a1bec2e5dc5d95d742941e6d3a5b548818c1b0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.260469 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.280442 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.285131 4744 scope.go:117] "RemoveContainer" containerID="8a8dd1d6f9777ebaa902f10e5b908ba2c65e7651dfcc963db3021e84fe0c725a" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.344108 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" path="/var/lib/kubelet/pods/ebb6d047-293a-42e1-9dd2-c35a736f93ea/volumes" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345024 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 09:11:32 crc kubenswrapper[4744]: E1201 09:11:32.345331 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-listener" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345346 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-listener" Dec 01 09:11:32 crc kubenswrapper[4744]: E1201 09:11:32.345363 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-evaluator" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345369 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-evaluator" Dec 01 09:11:32 crc kubenswrapper[4744]: E1201 09:11:32.345376 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-api" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345382 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-api" Dec 01 09:11:32 crc kubenswrapper[4744]: E1201 09:11:32.345391 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="extract-content" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345396 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="extract-content" Dec 01 09:11:32 crc kubenswrapper[4744]: E1201 09:11:32.345427 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="registry-server" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345433 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="registry-server" Dec 01 09:11:32 crc kubenswrapper[4744]: E1201 09:11:32.345446 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-notifier" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345451 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-notifier" Dec 01 09:11:32 crc kubenswrapper[4744]: E1201 09:11:32.345462 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="extract-utilities" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345468 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="extract-utilities" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345701 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-evaluator" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345722 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-listener" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345730 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d20c0c-e648-4e45-bab4-22a808ff3089" containerName="registry-server" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345738 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-notifier" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.345754 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb6d047-293a-42e1-9dd2-c35a736f93ea" containerName="aodh-api" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.347783 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.347869 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.358096 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.358267 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-h64d7" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.358379 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.358630 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.358755 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.389471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-public-tls-certs\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.389557 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-config-data\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.389774 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-scripts\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.389805 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.389925 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb6px\" (UniqueName: \"kubernetes.io/projected/c799ae29-585e-4e74-b14a-b60688ba13c0-kube-api-access-wb6px\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.390021 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-internal-tls-certs\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.400871 4744 scope.go:117] "RemoveContainer" containerID="3073308ea771cc8aa2f12ffa6d73e302ce7684c382d0bf167d592bb0b9ff25ee" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.491368 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb6px\" (UniqueName: \"kubernetes.io/projected/c799ae29-585e-4e74-b14a-b60688ba13c0-kube-api-access-wb6px\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.491435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-internal-tls-certs\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.491499 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-public-tls-certs\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.491549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-config-data\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.491627 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-scripts\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.491650 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.495862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-public-tls-certs\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.496984 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-config-data\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.499287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.510222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-internal-tls-certs\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.514706 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb6px\" (UniqueName: \"kubernetes.io/projected/c799ae29-585e-4e74-b14a-b60688ba13c0-kube-api-access-wb6px\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.526846 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-scripts\") pod \"aodh-0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " pod="openstack/aodh-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.558334 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.558393 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.560890 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:32 crc kubenswrapper[4744]: I1201 09:11:32.707295 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:11:33 crc kubenswrapper[4744]: W1201 09:11:33.181063 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc799ae29_585e_4e74_b14a_b60688ba13c0.slice/crio-1520c3a3ba72de9b78cc9b8c032f16d3afe1c60ad60b366160d5d54796dcf53b WatchSource:0}: Error finding container 1520c3a3ba72de9b78cc9b8c032f16d3afe1c60ad60b366160d5d54796dcf53b: Status 404 returned error can't find the container with id 1520c3a3ba72de9b78cc9b8c032f16d3afe1c60ad60b366160d5d54796dcf53b Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.181094 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.186472 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.814469 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vg8dk"] Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.817145 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.830678 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vg8dk"] Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.917169 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s877q\" (UniqueName: \"kubernetes.io/projected/d337f486-d205-420b-a648-c5d352adc56e-kube-api-access-s877q\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.917286 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-catalog-content\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:33 crc kubenswrapper[4744]: I1201 09:11:33.917312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-utilities\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.018939 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s877q\" (UniqueName: \"kubernetes.io/projected/d337f486-d205-420b-a648-c5d352adc56e-kube-api-access-s877q\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.019081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-catalog-content\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.019107 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-utilities\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.019786 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-catalog-content\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.019836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-utilities\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.045418 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s877q\" (UniqueName: \"kubernetes.io/projected/d337f486-d205-420b-a648-c5d352adc56e-kube-api-access-s877q\") pod \"community-operators-vg8dk\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.153082 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.195203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerStarted","Data":"1520c3a3ba72de9b78cc9b8c032f16d3afe1c60ad60b366160d5d54796dcf53b"} Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.709685 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vg8dk"] Dec 01 09:11:34 crc kubenswrapper[4744]: W1201 09:11:34.714127 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice/crio-e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100 WatchSource:0}: Error finding container e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100: Status 404 returned error can't find the container with id e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100 Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.729217 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.729448 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" containerName="openstackclient" containerID="cri-o://206eec330034b3141e9e3e9d46ba68368d3d0f9a7ae83fe695fc5810d8438667" gracePeriod=2 Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.742013 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.783271 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: E1201 09:11:34.786162 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" containerName="openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.786199 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" containerName="openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.789922 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" containerName="openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.793756 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.800056 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="210505c2-0d3f-4c88-8964-6c1194a43f47" podUID="55446f52-3cf3-4ac2-9ad1-7f35c5508697" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.807627 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" podUID="55446f52-3cf3-4ac2-9ad1-7f35c5508697" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.826536 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.846467 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.861623 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.880748 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.882454 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.892637 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:34 crc kubenswrapper[4744]: E1201 09:11:34.912867 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-vnv6l openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="210505c2-0d3f-4c88-8964-6c1194a43f47" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.959567 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnv6l\" (UniqueName: \"kubernetes.io/projected/210505c2-0d3f-4c88-8964-6c1194a43f47-kube-api-access-vnv6l\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.959642 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config-secret\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.959711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-combined-ca-bundle\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:34 crc kubenswrapper[4744]: I1201 09:11:34.959747 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.064811 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config-secret\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.065196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-combined-ca-bundle\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.065428 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72zxx\" (UniqueName: \"kubernetes.io/projected/55446f52-3cf3-4ac2-9ad1-7f35c5508697-kube-api-access-72zxx\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.065534 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55446f52-3cf3-4ac2-9ad1-7f35c5508697-combined-ca-bundle\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.065627 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.065744 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55446f52-3cf3-4ac2-9ad1-7f35c5508697-openstack-config-secret\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.065831 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55446f52-3cf3-4ac2-9ad1-7f35c5508697-openstack-config\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.065973 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnv6l\" (UniqueName: \"kubernetes.io/projected/210505c2-0d3f-4c88-8964-6c1194a43f47-kube-api-access-vnv6l\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.067135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.081328 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config-secret\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.081853 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-combined-ca-bundle\") pod \"openstackclient\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: E1201 09:11:35.082024 4744 projected.go:194] Error preparing data for projected volume kube-api-access-vnv6l for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (210505c2-0d3f-4c88-8964-6c1194a43f47) does not match the UID in record. The object might have been deleted and then recreated Dec 01 09:11:35 crc kubenswrapper[4744]: E1201 09:11:35.082068 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/210505c2-0d3f-4c88-8964-6c1194a43f47-kube-api-access-vnv6l podName:210505c2-0d3f-4c88-8964-6c1194a43f47 nodeName:}" failed. No retries permitted until 2025-12-01 09:11:35.582052223 +0000 UTC m=+3227.571110144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vnv6l" (UniqueName: "kubernetes.io/projected/210505c2-0d3f-4c88-8964-6c1194a43f47-kube-api-access-vnv6l") pod "openstackclient" (UID: "210505c2-0d3f-4c88-8964-6c1194a43f47") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (210505c2-0d3f-4c88-8964-6c1194a43f47) does not match the UID in record. The object might have been deleted and then recreated Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.167517 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55446f52-3cf3-4ac2-9ad1-7f35c5508697-openstack-config-secret\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.167838 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55446f52-3cf3-4ac2-9ad1-7f35c5508697-openstack-config\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.167977 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72zxx\" (UniqueName: \"kubernetes.io/projected/55446f52-3cf3-4ac2-9ad1-7f35c5508697-kube-api-access-72zxx\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.168000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55446f52-3cf3-4ac2-9ad1-7f35c5508697-combined-ca-bundle\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.168921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/55446f52-3cf3-4ac2-9ad1-7f35c5508697-openstack-config\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.176565 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55446f52-3cf3-4ac2-9ad1-7f35c5508697-combined-ca-bundle\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.195021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/55446f52-3cf3-4ac2-9ad1-7f35c5508697-openstack-config-secret\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.207103 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerStarted","Data":"7936b397970b6797f17a5d154fd5ffb635014f8fa2c12582ca7bf56022eb739a"} Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.208545 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72zxx\" (UniqueName: \"kubernetes.io/projected/55446f52-3cf3-4ac2-9ad1-7f35c5508697-kube-api-access-72zxx\") pod \"openstackclient\" (UID: \"55446f52-3cf3-4ac2-9ad1-7f35c5508697\") " pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.208924 4744 generic.go:334] "Generic (PLEG): container finished" podID="d337f486-d205-420b-a648-c5d352adc56e" containerID="4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e" exitCode=0 Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.208999 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.210696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vg8dk" event={"ID":"d337f486-d205-420b-a648-c5d352adc56e","Type":"ContainerDied","Data":"4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e"} Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.210733 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vg8dk" event={"ID":"d337f486-d205-420b-a648-c5d352adc56e","Type":"ContainerStarted","Data":"e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100"} Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.225443 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.234623 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.247789 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="210505c2-0d3f-4c88-8964-6c1194a43f47" podUID="55446f52-3cf3-4ac2-9ad1-7f35c5508697" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.307999 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.377615 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config-secret\") pod \"210505c2-0d3f-4c88-8964-6c1194a43f47\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.377794 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-combined-ca-bundle\") pod \"210505c2-0d3f-4c88-8964-6c1194a43f47\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.377972 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config\") pod \"210505c2-0d3f-4c88-8964-6c1194a43f47\" (UID: \"210505c2-0d3f-4c88-8964-6c1194a43f47\") " Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.378658 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnv6l\" (UniqueName: \"kubernetes.io/projected/210505c2-0d3f-4c88-8964-6c1194a43f47-kube-api-access-vnv6l\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.380326 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "210505c2-0d3f-4c88-8964-6c1194a43f47" (UID: "210505c2-0d3f-4c88-8964-6c1194a43f47"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.383380 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "210505c2-0d3f-4c88-8964-6c1194a43f47" (UID: "210505c2-0d3f-4c88-8964-6c1194a43f47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.391750 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "210505c2-0d3f-4c88-8964-6c1194a43f47" (UID: "210505c2-0d3f-4c88-8964-6c1194a43f47"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.481116 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.481522 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.481535 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/210505c2-0d3f-4c88-8964-6c1194a43f47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:35 crc kubenswrapper[4744]: I1201 09:11:35.888052 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 09:11:35 crc kubenswrapper[4744]: W1201 09:11:35.888835 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55446f52_3cf3_4ac2_9ad1_7f35c5508697.slice/crio-d4549cdd2cf4555073df2a6839430d934c547a85b4fc011c479baa16010698b6 WatchSource:0}: Error finding container d4549cdd2cf4555073df2a6839430d934c547a85b4fc011c479baa16010698b6: Status 404 returned error can't find the container with id d4549cdd2cf4555073df2a6839430d934c547a85b4fc011c479baa16010698b6 Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.230930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"55446f52-3cf3-4ac2-9ad1-7f35c5508697","Type":"ContainerStarted","Data":"65ef97f42694b35d63f120f64e8eda1547b2ba66a99036dd6fcca5fd47e3e27d"} Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.231252 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"55446f52-3cf3-4ac2-9ad1-7f35c5508697","Type":"ContainerStarted","Data":"d4549cdd2cf4555073df2a6839430d934c547a85b4fc011c479baa16010698b6"} Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.241026 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.242770 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerStarted","Data":"bdae28adaecc5affb55cf36ad90b8828a68b44e2c8377584c9afe6d192dfe71b"} Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.248358 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.2483181979999998 podStartE2EDuration="2.248318198s" podCreationTimestamp="2025-12-01 09:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:11:36.245394695 +0000 UTC m=+3228.234452616" watchObservedRunningTime="2025-12-01 09:11:36.248318198 +0000 UTC m=+3228.237376129" Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.250977 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="210505c2-0d3f-4c88-8964-6c1194a43f47" podUID="55446f52-3cf3-4ac2-9ad1-7f35c5508697" Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.259558 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="210505c2-0d3f-4c88-8964-6c1194a43f47" podUID="55446f52-3cf3-4ac2-9ad1-7f35c5508697" Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.307213 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210505c2-0d3f-4c88-8964-6c1194a43f47" path="/var/lib/kubelet/pods/210505c2-0d3f-4c88-8964-6c1194a43f47/volumes" Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.532005 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.532633 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="thanos-sidecar" containerID="cri-o://e4906d1a18342d363d9e8f79ec957a8c90ccebb1d1fe52fed2b03d4657ca030c" gracePeriod=600 Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.532657 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="config-reloader" containerID="cri-o://c2997a70a5b15362f5b693221990bd55322d1ac5e2ca0352d4f1905094c72211" gracePeriod=600 Dec 01 09:11:36 crc kubenswrapper[4744]: I1201 09:11:36.532593 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="prometheus" containerID="cri-o://2dd49f491993313e4adc9252a560e841527776d85de0248ec2415b1e213e27eb" gracePeriod=600 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.301025 4744 generic.go:334] "Generic (PLEG): container finished" podID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerID="e4906d1a18342d363d9e8f79ec957a8c90ccebb1d1fe52fed2b03d4657ca030c" exitCode=0 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.301628 4744 generic.go:334] "Generic (PLEG): container finished" podID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerID="2dd49f491993313e4adc9252a560e841527776d85de0248ec2415b1e213e27eb" exitCode=0 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.301258 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerDied","Data":"e4906d1a18342d363d9e8f79ec957a8c90ccebb1d1fe52fed2b03d4657ca030c"} Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.301820 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerDied","Data":"2dd49f491993313e4adc9252a560e841527776d85de0248ec2415b1e213e27eb"} Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.312026 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerStarted","Data":"53d6458f26c9c081a22ef3952715c3e4d9e414b695f844190f5415dbd1d9286f"} Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.312086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerStarted","Data":"82151126331ffae6b10fc71b1f80d14226c5db02ca3540e9b0e91ecc0accf52b"} Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.312242 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-api" containerID="cri-o://7936b397970b6797f17a5d154fd5ffb635014f8fa2c12582ca7bf56022eb739a" gracePeriod=30 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.312833 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-listener" containerID="cri-o://53d6458f26c9c081a22ef3952715c3e4d9e414b695f844190f5415dbd1d9286f" gracePeriod=30 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.312917 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-notifier" containerID="cri-o://82151126331ffae6b10fc71b1f80d14226c5db02ca3540e9b0e91ecc0accf52b" gracePeriod=30 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.312966 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-evaluator" containerID="cri-o://bdae28adaecc5affb55cf36ad90b8828a68b44e2c8377584c9afe6d192dfe71b" gracePeriod=30 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.344998 4744 generic.go:334] "Generic (PLEG): container finished" podID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" containerID="206eec330034b3141e9e3e9d46ba68368d3d0f9a7ae83fe695fc5810d8438667" exitCode=137 Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.356816 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vg8dk" event={"ID":"d337f486-d205-420b-a648-c5d352adc56e","Type":"ContainerStarted","Data":"39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af"} Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.358391 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.638795773 podStartE2EDuration="5.358365854s" podCreationTimestamp="2025-12-01 09:11:32 +0000 UTC" firstStartedPulling="2025-12-01 09:11:33.183776331 +0000 UTC m=+3225.172834262" lastFinishedPulling="2025-12-01 09:11:36.903346422 +0000 UTC m=+3228.892404343" observedRunningTime="2025-12-01 09:11:37.339991041 +0000 UTC m=+3229.329048962" watchObservedRunningTime="2025-12-01 09:11:37.358365854 +0000 UTC m=+3229.347423775" Dec 01 09:11:37 crc kubenswrapper[4744]: I1201 09:11:37.558573 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.1.9:9090/-/ready\": dial tcp 10.217.1.9:9090: connect: connection refused" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.279575 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.384972 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-combined-ca-bundle\") pod \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.385046 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4jwc\" (UniqueName: \"kubernetes.io/projected/ac90703c-38e1-4f7b-be3b-9b36f5467aab-kube-api-access-b4jwc\") pod \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.385171 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config-secret\") pod \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.385213 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config\") pod \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\" (UID: \"ac90703c-38e1-4f7b-be3b-9b36f5467aab\") " Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.411713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac90703c-38e1-4f7b-be3b-9b36f5467aab-kube-api-access-b4jwc" (OuterVolumeSpecName: "kube-api-access-b4jwc") pod "ac90703c-38e1-4f7b-be3b-9b36f5467aab" (UID: "ac90703c-38e1-4f7b-be3b-9b36f5467aab"). InnerVolumeSpecName "kube-api-access-b4jwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.425370 4744 generic.go:334] "Generic (PLEG): container finished" podID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerID="7936b397970b6797f17a5d154fd5ffb635014f8fa2c12582ca7bf56022eb739a" exitCode=0 Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.425462 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerDied","Data":"7936b397970b6797f17a5d154fd5ffb635014f8fa2c12582ca7bf56022eb739a"} Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.438791 4744 scope.go:117] "RemoveContainer" containerID="206eec330034b3141e9e3e9d46ba68368d3d0f9a7ae83fe695fc5810d8438667" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.439121 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.439123 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ac90703c-38e1-4f7b-be3b-9b36f5467aab" (UID: "ac90703c-38e1-4f7b-be3b-9b36f5467aab"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.443587 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac90703c-38e1-4f7b-be3b-9b36f5467aab" (UID: "ac90703c-38e1-4f7b-be3b-9b36f5467aab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.452386 4744 generic.go:334] "Generic (PLEG): container finished" podID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerID="c2997a70a5b15362f5b693221990bd55322d1ac5e2ca0352d4f1905094c72211" exitCode=0 Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.452516 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerDied","Data":"c2997a70a5b15362f5b693221990bd55322d1ac5e2ca0352d4f1905094c72211"} Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.487620 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.487658 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4jwc\" (UniqueName: \"kubernetes.io/projected/ac90703c-38e1-4f7b-be3b-9b36f5467aab-kube-api-access-b4jwc\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.487670 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.488947 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ac90703c-38e1-4f7b-be3b-9b36f5467aab" (UID: "ac90703c-38e1-4f7b-be3b-9b36f5467aab"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:38 crc kubenswrapper[4744]: I1201 09:11:38.589157 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ac90703c-38e1-4f7b-be3b-9b36f5467aab-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.116225 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.198772 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-thanos-prometheus-http-client-file\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.198934 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.199033 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.199105 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config-out\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.199168 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f40458d8-65cc-49fe-bbcb-16f0b4290d70-prometheus-metric-storage-rulefiles-0\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.199223 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88k8g\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-kube-api-access-88k8g\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.199257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-web-config\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.199299 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-tls-assets\") pod \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\" (UID: \"f40458d8-65cc-49fe-bbcb-16f0b4290d70\") " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.201749 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40458d8-65cc-49fe-bbcb-16f0b4290d70-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.204441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-kube-api-access-88k8g" (OuterVolumeSpecName: "kube-api-access-88k8g") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "kube-api-access-88k8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.208747 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.212234 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config" (OuterVolumeSpecName: "config") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.212243 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.212924 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config-out" (OuterVolumeSpecName: "config-out") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.219293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.249433 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-web-config" (OuterVolumeSpecName: "web-config") pod "f40458d8-65cc-49fe-bbcb-16f0b4290d70" (UID: "f40458d8-65cc-49fe-bbcb-16f0b4290d70"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302337 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302428 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302443 4744 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f40458d8-65cc-49fe-bbcb-16f0b4290d70-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302456 4744 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f40458d8-65cc-49fe-bbcb-16f0b4290d70-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302471 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88k8g\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-kube-api-access-88k8g\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302482 4744 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302491 4744 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f40458d8-65cc-49fe-bbcb-16f0b4290d70-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.302502 4744 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f40458d8-65cc-49fe-bbcb-16f0b4290d70-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.322149 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.404263 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.464880 4744 generic.go:334] "Generic (PLEG): container finished" podID="d337f486-d205-420b-a648-c5d352adc56e" containerID="39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af" exitCode=0 Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.464974 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vg8dk" event={"ID":"d337f486-d205-420b-a648-c5d352adc56e","Type":"ContainerDied","Data":"39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af"} Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.467892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f40458d8-65cc-49fe-bbcb-16f0b4290d70","Type":"ContainerDied","Data":"a5ced283d875c2fab6b59793fafeca56136afb9cf63400554096a76cd1488303"} Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.467955 4744 scope.go:117] "RemoveContainer" containerID="e4906d1a18342d363d9e8f79ec957a8c90ccebb1d1fe52fed2b03d4657ca030c" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.468114 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.499479 4744 generic.go:334] "Generic (PLEG): container finished" podID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerID="bdae28adaecc5affb55cf36ad90b8828a68b44e2c8377584c9afe6d192dfe71b" exitCode=0 Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.499560 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerDied","Data":"bdae28adaecc5affb55cf36ad90b8828a68b44e2c8377584c9afe6d192dfe71b"} Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.505471 4744 scope.go:117] "RemoveContainer" containerID="c2997a70a5b15362f5b693221990bd55322d1ac5e2ca0352d4f1905094c72211" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.530940 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.539308 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.551729 4744 scope.go:117] "RemoveContainer" containerID="2dd49f491993313e4adc9252a560e841527776d85de0248ec2415b1e213e27eb" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.583652 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:39 crc kubenswrapper[4744]: E1201 09:11:39.584170 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="init-config-reloader" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.584187 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="init-config-reloader" Dec 01 09:11:39 crc kubenswrapper[4744]: E1201 09:11:39.584205 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="config-reloader" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.584214 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="config-reloader" Dec 01 09:11:39 crc kubenswrapper[4744]: E1201 09:11:39.584240 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="prometheus" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.584249 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="prometheus" Dec 01 09:11:39 crc kubenswrapper[4744]: E1201 09:11:39.584261 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="thanos-sidecar" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.584268 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="thanos-sidecar" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.584560 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="prometheus" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.584593 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="config-reloader" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.584612 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" containerName="thanos-sidecar" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.586938 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.600154 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fdbml" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.600291 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.600353 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.601937 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.602102 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.602312 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607145 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6a50c1dd-5a51-4b73-8824-00669dbed793-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607196 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6a50c1dd-5a51-4b73-8824-00669dbed793-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607244 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607273 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607348 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607377 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607452 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607490 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-config\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607557 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607619 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt9s9\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-kube-api-access-qt9s9\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.607672 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.611871 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.615041 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.634377 4744 scope.go:117] "RemoveContainer" containerID="88ebf2a52c1ee892f132dcfdefc4dcdf0b44c1ac03e633702961be7a0fdbfd86" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709049 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-config\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709373 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709491 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt9s9\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-kube-api-access-qt9s9\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709533 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6a50c1dd-5a51-4b73-8824-00669dbed793-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709573 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6a50c1dd-5a51-4b73-8824-00669dbed793-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709605 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709651 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709707 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.709749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.710754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6a50c1dd-5a51-4b73-8824-00669dbed793-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.715762 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6a50c1dd-5a51-4b73-8824-00669dbed793-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.716580 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.717281 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.717490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-config\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.725153 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.725959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.728877 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.728985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.731471 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt9s9\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-kube-api-access-qt9s9\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.758048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:39 crc kubenswrapper[4744]: I1201 09:11:39.935147 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:11:40 crc kubenswrapper[4744]: I1201 09:11:40.286552 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:11:40 crc kubenswrapper[4744]: E1201 09:11:40.287088 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:11:40 crc kubenswrapper[4744]: I1201 09:11:40.296874 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac90703c-38e1-4f7b-be3b-9b36f5467aab" path="/var/lib/kubelet/pods/ac90703c-38e1-4f7b-be3b-9b36f5467aab/volumes" Dec 01 09:11:40 crc kubenswrapper[4744]: I1201 09:11:40.297821 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f40458d8-65cc-49fe-bbcb-16f0b4290d70" path="/var/lib/kubelet/pods/f40458d8-65cc-49fe-bbcb-16f0b4290d70/volumes" Dec 01 09:11:40 crc kubenswrapper[4744]: I1201 09:11:40.525727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vg8dk" event={"ID":"d337f486-d205-420b-a648-c5d352adc56e","Type":"ContainerStarted","Data":"ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db"} Dec 01 09:11:40 crc kubenswrapper[4744]: I1201 09:11:40.559904 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:11:40 crc kubenswrapper[4744]: I1201 09:11:40.562325 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vg8dk" podStartSLOduration=2.795712236 podStartE2EDuration="7.562314578s" podCreationTimestamp="2025-12-01 09:11:33 +0000 UTC" firstStartedPulling="2025-12-01 09:11:35.212027952 +0000 UTC m=+3227.201085873" lastFinishedPulling="2025-12-01 09:11:39.978630294 +0000 UTC m=+3231.967688215" observedRunningTime="2025-12-01 09:11:40.548387192 +0000 UTC m=+3232.537445123" watchObservedRunningTime="2025-12-01 09:11:40.562314578 +0000 UTC m=+3232.551372489" Dec 01 09:11:41 crc kubenswrapper[4744]: I1201 09:11:41.537667 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerStarted","Data":"60b848fa34845edf7e1236c7eb96ed2cb1c9f28761f85535e7044c4274abec6a"} Dec 01 09:11:44 crc kubenswrapper[4744]: I1201 09:11:44.153371 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:44 crc kubenswrapper[4744]: I1201 09:11:44.154350 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:44 crc kubenswrapper[4744]: I1201 09:11:44.198839 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:44 crc kubenswrapper[4744]: I1201 09:11:44.568021 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerStarted","Data":"32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e"} Dec 01 09:11:45 crc kubenswrapper[4744]: I1201 09:11:45.650293 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:45 crc kubenswrapper[4744]: I1201 09:11:45.700608 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vg8dk"] Dec 01 09:11:47 crc kubenswrapper[4744]: I1201 09:11:47.598680 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vg8dk" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="registry-server" containerID="cri-o://ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db" gracePeriod=2 Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.061472 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.180351 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-utilities\") pod \"d337f486-d205-420b-a648-c5d352adc56e\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.180842 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-catalog-content\") pod \"d337f486-d205-420b-a648-c5d352adc56e\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.180887 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s877q\" (UniqueName: \"kubernetes.io/projected/d337f486-d205-420b-a648-c5d352adc56e-kube-api-access-s877q\") pod \"d337f486-d205-420b-a648-c5d352adc56e\" (UID: \"d337f486-d205-420b-a648-c5d352adc56e\") " Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.181208 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-utilities" (OuterVolumeSpecName: "utilities") pod "d337f486-d205-420b-a648-c5d352adc56e" (UID: "d337f486-d205-420b-a648-c5d352adc56e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.181537 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.189716 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d337f486-d205-420b-a648-c5d352adc56e-kube-api-access-s877q" (OuterVolumeSpecName: "kube-api-access-s877q") pod "d337f486-d205-420b-a648-c5d352adc56e" (UID: "d337f486-d205-420b-a648-c5d352adc56e"). InnerVolumeSpecName "kube-api-access-s877q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.235814 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d337f486-d205-420b-a648-c5d352adc56e" (UID: "d337f486-d205-420b-a648-c5d352adc56e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.283074 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d337f486-d205-420b-a648-c5d352adc56e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.283128 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s877q\" (UniqueName: \"kubernetes.io/projected/d337f486-d205-420b-a648-c5d352adc56e-kube-api-access-s877q\") on node \"crc\" DevicePath \"\"" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.611378 4744 generic.go:334] "Generic (PLEG): container finished" podID="d337f486-d205-420b-a648-c5d352adc56e" containerID="ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db" exitCode=0 Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.611447 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vg8dk" event={"ID":"d337f486-d205-420b-a648-c5d352adc56e","Type":"ContainerDied","Data":"ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db"} Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.611479 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vg8dk" event={"ID":"d337f486-d205-420b-a648-c5d352adc56e","Type":"ContainerDied","Data":"e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100"} Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.611497 4744 scope.go:117] "RemoveContainer" containerID="ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.611496 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vg8dk" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.635667 4744 scope.go:117] "RemoveContainer" containerID="39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.645440 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vg8dk"] Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.656836 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vg8dk"] Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.663135 4744 scope.go:117] "RemoveContainer" containerID="4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.702056 4744 scope.go:117] "RemoveContainer" containerID="ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db" Dec 01 09:11:48 crc kubenswrapper[4744]: E1201 09:11:48.702530 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db\": container with ID starting with ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db not found: ID does not exist" containerID="ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.702653 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db"} err="failed to get container status \"ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db\": rpc error: code = NotFound desc = could not find container \"ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db\": container with ID starting with ffef6277dc5f32d57439645def3dd1db0a0eb0858142fb5350261e7b996dc8db not found: ID does not exist" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.702686 4744 scope.go:117] "RemoveContainer" containerID="39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af" Dec 01 09:11:48 crc kubenswrapper[4744]: E1201 09:11:48.703082 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af\": container with ID starting with 39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af not found: ID does not exist" containerID="39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.703114 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af"} err="failed to get container status \"39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af\": rpc error: code = NotFound desc = could not find container \"39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af\": container with ID starting with 39090cbdd8081e223b852baee9da68051e333264f6e4b5aa13ce6623993bf9af not found: ID does not exist" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.703129 4744 scope.go:117] "RemoveContainer" containerID="4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e" Dec 01 09:11:48 crc kubenswrapper[4744]: E1201 09:11:48.703618 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e\": container with ID starting with 4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e not found: ID does not exist" containerID="4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e" Dec 01 09:11:48 crc kubenswrapper[4744]: I1201 09:11:48.703666 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e"} err="failed to get container status \"4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e\": rpc error: code = NotFound desc = could not find container \"4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e\": container with ID starting with 4cba3776ac830a0f879fdba982d5453eccf94beee166d38f41f1fe077302e59e not found: ID does not exist" Dec 01 09:11:50 crc kubenswrapper[4744]: I1201 09:11:50.305507 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d337f486-d205-420b-a648-c5d352adc56e" path="/var/lib/kubelet/pods/d337f486-d205-420b-a648-c5d352adc56e/volumes" Dec 01 09:11:51 crc kubenswrapper[4744]: I1201 09:11:51.640584 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerID="32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e" exitCode=0 Dec 01 09:11:51 crc kubenswrapper[4744]: I1201 09:11:51.640652 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerDied","Data":"32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e"} Dec 01 09:11:52 crc kubenswrapper[4744]: I1201 09:11:52.651350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerStarted","Data":"65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a"} Dec 01 09:11:55 crc kubenswrapper[4744]: I1201 09:11:55.285573 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:11:55 crc kubenswrapper[4744]: E1201 09:11:55.286831 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:11:55 crc kubenswrapper[4744]: I1201 09:11:55.683816 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerStarted","Data":"b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a"} Dec 01 09:11:55 crc kubenswrapper[4744]: I1201 09:11:55.684121 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerStarted","Data":"8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9"} Dec 01 09:11:55 crc kubenswrapper[4744]: I1201 09:11:55.721575 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.721544556 podStartE2EDuration="16.721544556s" podCreationTimestamp="2025-12-01 09:11:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:11:55.710195953 +0000 UTC m=+3247.699253884" watchObservedRunningTime="2025-12-01 09:11:55.721544556 +0000 UTC m=+3247.710602477" Dec 01 09:11:59 crc kubenswrapper[4744]: I1201 09:11:59.936302 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 09:12:07 crc kubenswrapper[4744]: I1201 09:12:07.285300 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:12:07 crc kubenswrapper[4744]: E1201 09:12:07.286172 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:12:07 crc kubenswrapper[4744]: I1201 09:12:07.798496 4744 generic.go:334] "Generic (PLEG): container finished" podID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerID="53d6458f26c9c081a22ef3952715c3e4d9e414b695f844190f5415dbd1d9286f" exitCode=137 Dec 01 09:12:07 crc kubenswrapper[4744]: I1201 09:12:07.798747 4744 generic.go:334] "Generic (PLEG): container finished" podID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerID="82151126331ffae6b10fc71b1f80d14226c5db02ca3540e9b0e91ecc0accf52b" exitCode=137 Dec 01 09:12:07 crc kubenswrapper[4744]: I1201 09:12:07.798768 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerDied","Data":"53d6458f26c9c081a22ef3952715c3e4d9e414b695f844190f5415dbd1d9286f"} Dec 01 09:12:07 crc kubenswrapper[4744]: I1201 09:12:07.798793 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerDied","Data":"82151126331ffae6b10fc71b1f80d14226c5db02ca3540e9b0e91ecc0accf52b"} Dec 01 09:12:07 crc kubenswrapper[4744]: E1201 09:12:07.799775 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc799ae29_585e_4e74_b14a_b60688ba13c0.slice/crio-conmon-82151126331ffae6b10fc71b1f80d14226c5db02ca3540e9b0e91ecc0accf52b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice/crio-e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100\": RecentStats: unable to find data in memory cache]" Dec 01 09:12:07 crc kubenswrapper[4744]: I1201 09:12:07.924916 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.181618 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-config-data\") pod \"c799ae29-585e-4e74-b14a-b60688ba13c0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.181660 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-public-tls-certs\") pod \"c799ae29-585e-4e74-b14a-b60688ba13c0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.181701 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-combined-ca-bundle\") pod \"c799ae29-585e-4e74-b14a-b60688ba13c0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.181818 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-internal-tls-certs\") pod \"c799ae29-585e-4e74-b14a-b60688ba13c0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.181895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-scripts\") pod \"c799ae29-585e-4e74-b14a-b60688ba13c0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.181948 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb6px\" (UniqueName: \"kubernetes.io/projected/c799ae29-585e-4e74-b14a-b60688ba13c0-kube-api-access-wb6px\") pod \"c799ae29-585e-4e74-b14a-b60688ba13c0\" (UID: \"c799ae29-585e-4e74-b14a-b60688ba13c0\") " Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.188619 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-scripts" (OuterVolumeSpecName: "scripts") pod "c799ae29-585e-4e74-b14a-b60688ba13c0" (UID: "c799ae29-585e-4e74-b14a-b60688ba13c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.197757 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c799ae29-585e-4e74-b14a-b60688ba13c0-kube-api-access-wb6px" (OuterVolumeSpecName: "kube-api-access-wb6px") pod "c799ae29-585e-4e74-b14a-b60688ba13c0" (UID: "c799ae29-585e-4e74-b14a-b60688ba13c0"). InnerVolumeSpecName "kube-api-access-wb6px". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.245207 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c799ae29-585e-4e74-b14a-b60688ba13c0" (UID: "c799ae29-585e-4e74-b14a-b60688ba13c0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.265357 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c799ae29-585e-4e74-b14a-b60688ba13c0" (UID: "c799ae29-585e-4e74-b14a-b60688ba13c0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.284127 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.284162 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb6px\" (UniqueName: \"kubernetes.io/projected/c799ae29-585e-4e74-b14a-b60688ba13c0-kube-api-access-wb6px\") on node \"crc\" DevicePath \"\"" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.284174 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.284182 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.312097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-config-data" (OuterVolumeSpecName: "config-data") pod "c799ae29-585e-4e74-b14a-b60688ba13c0" (UID: "c799ae29-585e-4e74-b14a-b60688ba13c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.318039 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c799ae29-585e-4e74-b14a-b60688ba13c0" (UID: "c799ae29-585e-4e74-b14a-b60688ba13c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.404775 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.404892 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c799ae29-585e-4e74-b14a-b60688ba13c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.810749 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c799ae29-585e-4e74-b14a-b60688ba13c0","Type":"ContainerDied","Data":"1520c3a3ba72de9b78cc9b8c032f16d3afe1c60ad60b366160d5d54796dcf53b"} Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.810828 4744 scope.go:117] "RemoveContainer" containerID="53d6458f26c9c081a22ef3952715c3e4d9e414b695f844190f5415dbd1d9286f" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.810826 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.837354 4744 scope.go:117] "RemoveContainer" containerID="82151126331ffae6b10fc71b1f80d14226c5db02ca3540e9b0e91ecc0accf52b" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.850730 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.859383 4744 scope.go:117] "RemoveContainer" containerID="bdae28adaecc5affb55cf36ad90b8828a68b44e2c8377584c9afe6d192dfe71b" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.862902 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.882307 4744 scope.go:117] "RemoveContainer" containerID="7936b397970b6797f17a5d154fd5ffb635014f8fa2c12582ca7bf56022eb739a" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901152 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 09:12:08 crc kubenswrapper[4744]: E1201 09:12:08.901700 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="extract-utilities" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901723 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="extract-utilities" Dec 01 09:12:08 crc kubenswrapper[4744]: E1201 09:12:08.901748 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-evaluator" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901756 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-evaluator" Dec 01 09:12:08 crc kubenswrapper[4744]: E1201 09:12:08.901770 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="registry-server" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901777 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="registry-server" Dec 01 09:12:08 crc kubenswrapper[4744]: E1201 09:12:08.901788 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-api" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901795 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-api" Dec 01 09:12:08 crc kubenswrapper[4744]: E1201 09:12:08.901808 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-notifier" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901818 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-notifier" Dec 01 09:12:08 crc kubenswrapper[4744]: E1201 09:12:08.901836 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="extract-content" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901844 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="extract-content" Dec 01 09:12:08 crc kubenswrapper[4744]: E1201 09:12:08.901860 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-listener" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.901866 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-listener" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.902085 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-listener" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.902107 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-notifier" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.902131 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-evaluator" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.902147 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" containerName="aodh-api" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.902162 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d337f486-d205-420b-a648-c5d352adc56e" containerName="registry-server" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.904389 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.908793 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.909272 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.909462 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-h64d7" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.909706 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.917524 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.918197 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-internal-tls-certs\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.918370 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-config-data\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.918490 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk69g\" (UniqueName: \"kubernetes.io/projected/39cfebb3-d0da-4535-8c1c-81d1db536ae1-kube-api-access-hk69g\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.918669 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-public-tls-certs\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.918705 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-scripts\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.918741 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:08 crc kubenswrapper[4744]: I1201 09:12:08.920301 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.021182 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-config-data\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.021256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk69g\" (UniqueName: \"kubernetes.io/projected/39cfebb3-d0da-4535-8c1c-81d1db536ae1-kube-api-access-hk69g\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.021344 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-public-tls-certs\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.021376 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-scripts\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.021451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.021596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-internal-tls-certs\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.025689 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-public-tls-certs\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.025974 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.026123 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-config-data\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.029282 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-scripts\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.030320 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-internal-tls-certs\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.042713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk69g\" (UniqueName: \"kubernetes.io/projected/39cfebb3-d0da-4535-8c1c-81d1db536ae1-kube-api-access-hk69g\") pod \"aodh-0\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.229091 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.658134 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:12:09 crc kubenswrapper[4744]: W1201 09:12:09.668592 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39cfebb3_d0da_4535_8c1c_81d1db536ae1.slice/crio-0b2c4c5899bb4c2ad2568b660d8398854495de2e00e6740165d9e446eaa105b1 WatchSource:0}: Error finding container 0b2c4c5899bb4c2ad2568b660d8398854495de2e00e6740165d9e446eaa105b1: Status 404 returned error can't find the container with id 0b2c4c5899bb4c2ad2568b660d8398854495de2e00e6740165d9e446eaa105b1 Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.822919 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerStarted","Data":"0b2c4c5899bb4c2ad2568b660d8398854495de2e00e6740165d9e446eaa105b1"} Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.935928 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 09:12:09 crc kubenswrapper[4744]: I1201 09:12:09.942636 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 09:12:10 crc kubenswrapper[4744]: I1201 09:12:10.300365 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c799ae29-585e-4e74-b14a-b60688ba13c0" path="/var/lib/kubelet/pods/c799ae29-585e-4e74-b14a-b60688ba13c0/volumes" Dec 01 09:12:10 crc kubenswrapper[4744]: I1201 09:12:10.834238 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerStarted","Data":"f71583f8c36937b9bbdd8512fbd021598926093cf2fa218a36112060c388cec4"} Dec 01 09:12:10 crc kubenswrapper[4744]: I1201 09:12:10.839808 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 09:12:11 crc kubenswrapper[4744]: I1201 09:12:11.864134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerStarted","Data":"ead96cef32b03beee73b1a60f8ebcaa35e82a3012cfffaff3937842f55c57c48"} Dec 01 09:12:13 crc kubenswrapper[4744]: I1201 09:12:13.894524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerStarted","Data":"b01f3c5f3468b23fef58259a77487f183675f1750e4d53b954e01f8de225604f"} Dec 01 09:12:13 crc kubenswrapper[4744]: I1201 09:12:13.895255 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerStarted","Data":"5093f26ba56fbf273d06c606f3d7da2514fd550cf20024acd5ae8ca80195905e"} Dec 01 09:12:13 crc kubenswrapper[4744]: I1201 09:12:13.918588 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.211026224 podStartE2EDuration="5.918559902s" podCreationTimestamp="2025-12-01 09:12:08 +0000 UTC" firstStartedPulling="2025-12-01 09:12:09.672631489 +0000 UTC m=+3261.661689420" lastFinishedPulling="2025-12-01 09:12:13.380165177 +0000 UTC m=+3265.369223098" observedRunningTime="2025-12-01 09:12:13.912580841 +0000 UTC m=+3265.901638782" watchObservedRunningTime="2025-12-01 09:12:13.918559902 +0000 UTC m=+3265.907617823" Dec 01 09:12:18 crc kubenswrapper[4744]: E1201 09:12:18.051009 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice/crio-e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice\": RecentStats: unable to find data in memory cache]" Dec 01 09:12:21 crc kubenswrapper[4744]: I1201 09:12:21.284912 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:12:21 crc kubenswrapper[4744]: E1201 09:12:21.285773 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:12:28 crc kubenswrapper[4744]: E1201 09:12:28.302991 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice/crio-e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100\": RecentStats: unable to find data in memory cache]" Dec 01 09:12:34 crc kubenswrapper[4744]: I1201 09:12:34.286014 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:12:34 crc kubenswrapper[4744]: E1201 09:12:34.287879 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:12:38 crc kubenswrapper[4744]: E1201 09:12:38.607554 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice/crio-e80ad326ce31cbd9d0a830fc2f14a232e3ffaead81c8714787d8f99df5bc2100\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd337f486_d205_420b_a648_c5d352adc56e.slice\": RecentStats: unable to find data in memory cache]" Dec 01 09:12:45 crc kubenswrapper[4744]: I1201 09:12:45.285401 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:12:45 crc kubenswrapper[4744]: E1201 09:12:45.286171 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:13:00 crc kubenswrapper[4744]: I1201 09:13:00.291541 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:13:01 crc kubenswrapper[4744]: I1201 09:13:01.338221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"c19ec233f541084953b250fd9ef16a3aa9905562e8421df21f2178ae31c491d1"} Dec 01 09:13:38 crc kubenswrapper[4744]: I1201 09:13:38.738738 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.536291 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.537043 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="config-reloader" containerID="cri-o://8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9" gracePeriod=600 Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.537062 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="thanos-sidecar" containerID="cri-o://b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a" gracePeriod=600 Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.536994 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="prometheus" containerID="cri-o://65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a" gracePeriod=600 Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.761001 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerID="b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a" exitCode=0 Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.761030 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerID="65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a" exitCode=0 Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.761048 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerDied","Data":"b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a"} Dec 01 09:13:40 crc kubenswrapper[4744]: I1201 09:13:40.761072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerDied","Data":"65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a"} Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.539220 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658377 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt9s9\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-kube-api-access-qt9s9\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658552 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658593 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658625 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6a50c1dd-5a51-4b73-8824-00669dbed793-config-out\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658687 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6a50c1dd-5a51-4b73-8824-00669dbed793-prometheus-metric-storage-rulefiles-0\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-secret-combined-ca-bundle\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658785 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658858 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-thanos-prometheus-http-client-file\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.658977 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-config\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.659018 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-tls-assets\") pod \"6a50c1dd-5a51-4b73-8824-00669dbed793\" (UID: \"6a50c1dd-5a51-4b73-8824-00669dbed793\") " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.659935 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a50c1dd-5a51-4b73-8824-00669dbed793-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.670620 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.678030 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-config" (OuterVolumeSpecName: "config") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.680542 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.681815 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a50c1dd-5a51-4b73-8824-00669dbed793-config-out" (OuterVolumeSpecName: "config-out") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.683675 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.690625 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-kube-api-access-qt9s9" (OuterVolumeSpecName: "kube-api-access-qt9s9") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "kube-api-access-qt9s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.702126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.720596 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.723659 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762090 4744 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762129 4744 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762143 4744 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762155 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762165 4744 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762177 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt9s9\" (UniqueName: \"kubernetes.io/projected/6a50c1dd-5a51-4b73-8824-00669dbed793-kube-api-access-qt9s9\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762204 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762215 4744 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6a50c1dd-5a51-4b73-8824-00669dbed793-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762227 4744 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6a50c1dd-5a51-4b73-8824-00669dbed793-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.762238 4744 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.775000 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerID="8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9" exitCode=0 Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.775052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerDied","Data":"8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9"} Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.775092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6a50c1dd-5a51-4b73-8824-00669dbed793","Type":"ContainerDied","Data":"60b848fa34845edf7e1236c7eb96ed2cb1c9f28761f85535e7044c4274abec6a"} Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.775094 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.775113 4744 scope.go:117] "RemoveContainer" containerID="b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.789754 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.831742 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config" (OuterVolumeSpecName: "web-config") pod "6a50c1dd-5a51-4b73-8824-00669dbed793" (UID: "6a50c1dd-5a51-4b73-8824-00669dbed793"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.863636 4744 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6a50c1dd-5a51-4b73-8824-00669dbed793-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.863676 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.888243 4744 scope.go:117] "RemoveContainer" containerID="8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.906052 4744 scope.go:117] "RemoveContainer" containerID="65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.924512 4744 scope.go:117] "RemoveContainer" containerID="32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.946889 4744 scope.go:117] "RemoveContainer" containerID="b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a" Dec 01 09:13:41 crc kubenswrapper[4744]: E1201 09:13:41.947711 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a\": container with ID starting with b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a not found: ID does not exist" containerID="b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.947749 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a"} err="failed to get container status \"b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a\": rpc error: code = NotFound desc = could not find container \"b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a\": container with ID starting with b6e3cd97536b252b238fb9bf5bba687886b10d173f4c532d488f29bda3fcf68a not found: ID does not exist" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.947777 4744 scope.go:117] "RemoveContainer" containerID="8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9" Dec 01 09:13:41 crc kubenswrapper[4744]: E1201 09:13:41.948101 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9\": container with ID starting with 8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9 not found: ID does not exist" containerID="8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.948161 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9"} err="failed to get container status \"8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9\": rpc error: code = NotFound desc = could not find container \"8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9\": container with ID starting with 8eb8fa7c313cab07541046ead8e925ef6cb4266324e4fd58bd32e15fe6accda9 not found: ID does not exist" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.948194 4744 scope.go:117] "RemoveContainer" containerID="65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a" Dec 01 09:13:41 crc kubenswrapper[4744]: E1201 09:13:41.948627 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a\": container with ID starting with 65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a not found: ID does not exist" containerID="65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.948667 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a"} err="failed to get container status \"65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a\": rpc error: code = NotFound desc = could not find container \"65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a\": container with ID starting with 65bed123af6867704da503067993bb28555cdaad32ca72e4d1067a677c98f37a not found: ID does not exist" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.948694 4744 scope.go:117] "RemoveContainer" containerID="32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e" Dec 01 09:13:41 crc kubenswrapper[4744]: E1201 09:13:41.949050 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e\": container with ID starting with 32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e not found: ID does not exist" containerID="32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e" Dec 01 09:13:41 crc kubenswrapper[4744]: I1201 09:13:41.949077 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e"} err="failed to get container status \"32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e\": rpc error: code = NotFound desc = could not find container \"32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e\": container with ID starting with 32d788680259c23c67a4ff1c55809087cd05fdd447a90da9d0bcbc221196c61e not found: ID does not exist" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.110192 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.119918 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.296359 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" path="/var/lib/kubelet/pods/6a50c1dd-5a51-4b73-8824-00669dbed793/volumes" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.852693 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:13:42 crc kubenswrapper[4744]: E1201 09:13:42.853088 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="config-reloader" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.853100 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="config-reloader" Dec 01 09:13:42 crc kubenswrapper[4744]: E1201 09:13:42.853121 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="prometheus" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.853127 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="prometheus" Dec 01 09:13:42 crc kubenswrapper[4744]: E1201 09:13:42.853136 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="init-config-reloader" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.853142 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="init-config-reloader" Dec 01 09:13:42 crc kubenswrapper[4744]: E1201 09:13:42.853166 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="thanos-sidecar" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.853173 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="thanos-sidecar" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.853383 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="config-reloader" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.853403 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="thanos-sidecar" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.853432 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a50c1dd-5a51-4b73-8824-00669dbed793" containerName="prometheus" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.855486 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.860438 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.863201 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.863267 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.863438 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fdbml" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.863832 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.863882 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.869027 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.874914 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.883561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.883620 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.883665 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.883690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.883729 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.883783 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.883819 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.884037 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.884062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9c79\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-kube-api-access-q9c79\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.884104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.884172 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986049 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986151 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986192 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986276 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9c79\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-kube-api-access-q9c79\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986332 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.986525 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.987541 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.991739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.991845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.991927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.991963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.992850 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.996721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:42 crc kubenswrapper[4744]: I1201 09:13:42.997628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:43 crc kubenswrapper[4744]: I1201 09:13:43.007458 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9c79\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-kube-api-access-q9c79\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:43 crc kubenswrapper[4744]: I1201 09:13:43.007579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:43 crc kubenswrapper[4744]: I1201 09:13:43.176187 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:13:43 crc kubenswrapper[4744]: I1201 09:13:43.652296 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:13:43 crc kubenswrapper[4744]: I1201 09:13:43.798421 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerStarted","Data":"7d61ba4b1118c979921f4f886981d6b2a271a7c66dbf09e7573ef53a79ceccbf"} Dec 01 09:13:47 crc kubenswrapper[4744]: I1201 09:13:47.850933 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerStarted","Data":"ef0ec917e4bf513772fac217a7b3b0f84f1b7f7e2d9ad2cce73654caf977aede"} Dec 01 09:13:54 crc kubenswrapper[4744]: I1201 09:13:54.918371 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerID="ef0ec917e4bf513772fac217a7b3b0f84f1b7f7e2d9ad2cce73654caf977aede" exitCode=0 Dec 01 09:13:54 crc kubenswrapper[4744]: I1201 09:13:54.918850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerDied","Data":"ef0ec917e4bf513772fac217a7b3b0f84f1b7f7e2d9ad2cce73654caf977aede"} Dec 01 09:13:55 crc kubenswrapper[4744]: I1201 09:13:55.929000 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerStarted","Data":"49586ba69df19a1e1d1f847eefac4279a25497755f28a7723ff03db86b58080d"} Dec 01 09:13:58 crc kubenswrapper[4744]: I1201 09:13:58.959733 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerStarted","Data":"c779f4c0285b53f8939ed894af7c1bdab911906297fa7af4846d96b0bfbb37c2"} Dec 01 09:13:59 crc kubenswrapper[4744]: I1201 09:13:59.973218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerStarted","Data":"d3b67176fdf366f4501f6d424406e7857f33ba75b8ef1cf4f99a71f902db2517"} Dec 01 09:14:00 crc kubenswrapper[4744]: I1201 09:14:00.002816 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.002794972 podStartE2EDuration="18.002794972s" podCreationTimestamp="2025-12-01 09:13:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:13:59.998222312 +0000 UTC m=+3371.987280243" watchObservedRunningTime="2025-12-01 09:14:00.002794972 +0000 UTC m=+3371.991852893" Dec 01 09:14:03 crc kubenswrapper[4744]: I1201 09:14:03.177248 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 09:14:13 crc kubenswrapper[4744]: I1201 09:14:13.177346 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 09:14:13 crc kubenswrapper[4744]: I1201 09:14:13.192258 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 09:14:14 crc kubenswrapper[4744]: I1201 09:14:14.118821 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.164345 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2"] Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.167186 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.169540 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.169795 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.180656 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2"] Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.284522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqg79\" (UniqueName: \"kubernetes.io/projected/f2376cd7-9853-4173-8313-3b0e9f346ac6-kube-api-access-qqg79\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.284660 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2376cd7-9853-4173-8313-3b0e9f346ac6-secret-volume\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.284703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2376cd7-9853-4173-8313-3b0e9f346ac6-config-volume\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.387032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2376cd7-9853-4173-8313-3b0e9f346ac6-secret-volume\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.387111 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2376cd7-9853-4173-8313-3b0e9f346ac6-config-volume\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.387265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqg79\" (UniqueName: \"kubernetes.io/projected/f2376cd7-9853-4173-8313-3b0e9f346ac6-kube-api-access-qqg79\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.388462 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2376cd7-9853-4173-8313-3b0e9f346ac6-config-volume\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.392635 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2376cd7-9853-4173-8313-3b0e9f346ac6-secret-volume\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.403566 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqg79\" (UniqueName: \"kubernetes.io/projected/f2376cd7-9853-4173-8313-3b0e9f346ac6-kube-api-access-qqg79\") pod \"collect-profiles-29409675-b4gx2\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.495607 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:00 crc kubenswrapper[4744]: I1201 09:15:00.961165 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2"] Dec 01 09:15:01 crc kubenswrapper[4744]: I1201 09:15:01.624194 4744 generic.go:334] "Generic (PLEG): container finished" podID="f2376cd7-9853-4173-8313-3b0e9f346ac6" containerID="be64e732c9d4f823e2e646529f8545b1a4a083bc28f1646ddb3b9b6b2d1418fb" exitCode=0 Dec 01 09:15:01 crc kubenswrapper[4744]: I1201 09:15:01.624445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" event={"ID":"f2376cd7-9853-4173-8313-3b0e9f346ac6","Type":"ContainerDied","Data":"be64e732c9d4f823e2e646529f8545b1a4a083bc28f1646ddb3b9b6b2d1418fb"} Dec 01 09:15:01 crc kubenswrapper[4744]: I1201 09:15:01.624539 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" event={"ID":"f2376cd7-9853-4173-8313-3b0e9f346ac6","Type":"ContainerStarted","Data":"3436ea3d144d434a53c9604bc152f14526b66b7e37f86b53c175150d457684f4"} Dec 01 09:15:02 crc kubenswrapper[4744]: I1201 09:15:02.967353 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.053512 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2376cd7-9853-4173-8313-3b0e9f346ac6-config-volume\") pod \"f2376cd7-9853-4173-8313-3b0e9f346ac6\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.053563 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2376cd7-9853-4173-8313-3b0e9f346ac6-secret-volume\") pod \"f2376cd7-9853-4173-8313-3b0e9f346ac6\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.053635 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqg79\" (UniqueName: \"kubernetes.io/projected/f2376cd7-9853-4173-8313-3b0e9f346ac6-kube-api-access-qqg79\") pod \"f2376cd7-9853-4173-8313-3b0e9f346ac6\" (UID: \"f2376cd7-9853-4173-8313-3b0e9f346ac6\") " Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.054571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2376cd7-9853-4173-8313-3b0e9f346ac6-config-volume" (OuterVolumeSpecName: "config-volume") pod "f2376cd7-9853-4173-8313-3b0e9f346ac6" (UID: "f2376cd7-9853-4173-8313-3b0e9f346ac6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.055026 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2376cd7-9853-4173-8313-3b0e9f346ac6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.062095 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2376cd7-9853-4173-8313-3b0e9f346ac6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f2376cd7-9853-4173-8313-3b0e9f346ac6" (UID: "f2376cd7-9853-4173-8313-3b0e9f346ac6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.062478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2376cd7-9853-4173-8313-3b0e9f346ac6-kube-api-access-qqg79" (OuterVolumeSpecName: "kube-api-access-qqg79") pod "f2376cd7-9853-4173-8313-3b0e9f346ac6" (UID: "f2376cd7-9853-4173-8313-3b0e9f346ac6"). InnerVolumeSpecName "kube-api-access-qqg79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.156729 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2376cd7-9853-4173-8313-3b0e9f346ac6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.156777 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqg79\" (UniqueName: \"kubernetes.io/projected/f2376cd7-9853-4173-8313-3b0e9f346ac6-kube-api-access-qqg79\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.646100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" event={"ID":"f2376cd7-9853-4173-8313-3b0e9f346ac6","Type":"ContainerDied","Data":"3436ea3d144d434a53c9604bc152f14526b66b7e37f86b53c175150d457684f4"} Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.646202 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3436ea3d144d434a53c9604bc152f14526b66b7e37f86b53c175150d457684f4" Dec 01 09:15:03 crc kubenswrapper[4744]: I1201 09:15:03.646310 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-b4gx2" Dec 01 09:15:04 crc kubenswrapper[4744]: I1201 09:15:04.060813 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt"] Dec 01 09:15:04 crc kubenswrapper[4744]: I1201 09:15:04.074243 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-mklbt"] Dec 01 09:15:04 crc kubenswrapper[4744]: I1201 09:15:04.329914 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="832ffd1d-24fd-4170-b513-b9381be00bce" path="/var/lib/kubelet/pods/832ffd1d-24fd-4170-b513-b9381be00bce/volumes" Dec 01 09:15:18 crc kubenswrapper[4744]: I1201 09:15:18.635056 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:15:18 crc kubenswrapper[4744]: I1201 09:15:18.635908 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:15:40 crc kubenswrapper[4744]: I1201 09:15:40.681614 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:15:41 crc kubenswrapper[4744]: I1201 09:15:41.999505 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:15:42 crc kubenswrapper[4744]: I1201 09:15:41.999834 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-api" containerID="cri-o://f71583f8c36937b9bbdd8512fbd021598926093cf2fa218a36112060c388cec4" gracePeriod=30 Dec 01 09:15:42 crc kubenswrapper[4744]: I1201 09:15:41.999901 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-listener" containerID="cri-o://b01f3c5f3468b23fef58259a77487f183675f1750e4d53b954e01f8de225604f" gracePeriod=30 Dec 01 09:15:42 crc kubenswrapper[4744]: I1201 09:15:41.999929 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-notifier" containerID="cri-o://5093f26ba56fbf273d06c606f3d7da2514fd550cf20024acd5ae8ca80195905e" gracePeriod=30 Dec 01 09:15:42 crc kubenswrapper[4744]: I1201 09:15:41.999974 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-evaluator" containerID="cri-o://ead96cef32b03beee73b1a60f8ebcaa35e82a3012cfffaff3937842f55c57c48" gracePeriod=30 Dec 01 09:15:43 crc kubenswrapper[4744]: I1201 09:15:43.013352 4744 generic.go:334] "Generic (PLEG): container finished" podID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerID="ead96cef32b03beee73b1a60f8ebcaa35e82a3012cfffaff3937842f55c57c48" exitCode=0 Dec 01 09:15:43 crc kubenswrapper[4744]: I1201 09:15:43.013689 4744 generic.go:334] "Generic (PLEG): container finished" podID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerID="f71583f8c36937b9bbdd8512fbd021598926093cf2fa218a36112060c388cec4" exitCode=0 Dec 01 09:15:43 crc kubenswrapper[4744]: I1201 09:15:43.013452 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerDied","Data":"ead96cef32b03beee73b1a60f8ebcaa35e82a3012cfffaff3937842f55c57c48"} Dec 01 09:15:43 crc kubenswrapper[4744]: I1201 09:15:43.013728 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerDied","Data":"f71583f8c36937b9bbdd8512fbd021598926093cf2fa218a36112060c388cec4"} Dec 01 09:15:48 crc kubenswrapper[4744]: I1201 09:15:48.634439 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:15:48 crc kubenswrapper[4744]: I1201 09:15:48.635060 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.085892 4744 generic.go:334] "Generic (PLEG): container finished" podID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerID="b01f3c5f3468b23fef58259a77487f183675f1750e4d53b954e01f8de225604f" exitCode=0 Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.086363 4744 generic.go:334] "Generic (PLEG): container finished" podID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerID="5093f26ba56fbf273d06c606f3d7da2514fd550cf20024acd5ae8ca80195905e" exitCode=0 Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.086044 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerDied","Data":"b01f3c5f3468b23fef58259a77487f183675f1750e4d53b954e01f8de225604f"} Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.086419 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerDied","Data":"5093f26ba56fbf273d06c606f3d7da2514fd550cf20024acd5ae8ca80195905e"} Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.368625 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.525683 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk69g\" (UniqueName: \"kubernetes.io/projected/39cfebb3-d0da-4535-8c1c-81d1db536ae1-kube-api-access-hk69g\") pod \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.525768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-internal-tls-certs\") pod \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.525821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-public-tls-certs\") pod \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.525899 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-scripts\") pod \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.525938 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-combined-ca-bundle\") pod \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.526116 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-config-data\") pod \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\" (UID: \"39cfebb3-d0da-4535-8c1c-81d1db536ae1\") " Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.531463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-scripts" (OuterVolumeSpecName: "scripts") pod "39cfebb3-d0da-4535-8c1c-81d1db536ae1" (UID: "39cfebb3-d0da-4535-8c1c-81d1db536ae1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.531469 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39cfebb3-d0da-4535-8c1c-81d1db536ae1-kube-api-access-hk69g" (OuterVolumeSpecName: "kube-api-access-hk69g") pod "39cfebb3-d0da-4535-8c1c-81d1db536ae1" (UID: "39cfebb3-d0da-4535-8c1c-81d1db536ae1"). InnerVolumeSpecName "kube-api-access-hk69g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.642106 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.642150 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk69g\" (UniqueName: \"kubernetes.io/projected/39cfebb3-d0da-4535-8c1c-81d1db536ae1-kube-api-access-hk69g\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.703585 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "39cfebb3-d0da-4535-8c1c-81d1db536ae1" (UID: "39cfebb3-d0da-4535-8c1c-81d1db536ae1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.732548 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-config-data" (OuterVolumeSpecName: "config-data") pod "39cfebb3-d0da-4535-8c1c-81d1db536ae1" (UID: "39cfebb3-d0da-4535-8c1c-81d1db536ae1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.742637 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "39cfebb3-d0da-4535-8c1c-81d1db536ae1" (UID: "39cfebb3-d0da-4535-8c1c-81d1db536ae1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.743909 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.743947 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.743959 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.749589 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39cfebb3-d0da-4535-8c1c-81d1db536ae1" (UID: "39cfebb3-d0da-4535-8c1c-81d1db536ae1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:51 crc kubenswrapper[4744]: I1201 09:15:51.846684 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cfebb3-d0da-4535-8c1c-81d1db536ae1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.096734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"39cfebb3-d0da-4535-8c1c-81d1db536ae1","Type":"ContainerDied","Data":"0b2c4c5899bb4c2ad2568b660d8398854495de2e00e6740165d9e446eaa105b1"} Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.096792 4744 scope.go:117] "RemoveContainer" containerID="b01f3c5f3468b23fef58259a77487f183675f1750e4d53b954e01f8de225604f" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.096804 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.136993 4744 scope.go:117] "RemoveContainer" containerID="5093f26ba56fbf273d06c606f3d7da2514fd550cf20024acd5ae8ca80195905e" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.148261 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.168483 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.188240 4744 scope.go:117] "RemoveContainer" containerID="ead96cef32b03beee73b1a60f8ebcaa35e82a3012cfffaff3937842f55c57c48" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.190866 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 09:15:52 crc kubenswrapper[4744]: E1201 09:15:52.191325 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-listener" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191344 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-listener" Dec 01 09:15:52 crc kubenswrapper[4744]: E1201 09:15:52.191363 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-api" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191372 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-api" Dec 01 09:15:52 crc kubenswrapper[4744]: E1201 09:15:52.191391 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-notifier" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191399 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-notifier" Dec 01 09:15:52 crc kubenswrapper[4744]: E1201 09:15:52.191434 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-evaluator" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191442 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-evaluator" Dec 01 09:15:52 crc kubenswrapper[4744]: E1201 09:15:52.191460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2376cd7-9853-4173-8313-3b0e9f346ac6" containerName="collect-profiles" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191467 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2376cd7-9853-4173-8313-3b0e9f346ac6" containerName="collect-profiles" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191704 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-evaluator" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191721 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2376cd7-9853-4173-8313-3b0e9f346ac6" containerName="collect-profiles" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191743 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-listener" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191759 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-notifier" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.191780 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" containerName="aodh-api" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.193628 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.196476 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.198768 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-h64d7" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.199047 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.199194 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.199393 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.222299 4744 scope.go:117] "RemoveContainer" containerID="f71583f8c36937b9bbdd8512fbd021598926093cf2fa218a36112060c388cec4" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.230929 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.261951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-internal-tls-certs\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.262013 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-scripts\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.262091 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-public-tls-certs\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.262146 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfrjk\" (UniqueName: \"kubernetes.io/projected/00e7d3a7-76c8-4657-9168-5086d95c429a-kube-api-access-mfrjk\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.262219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.262270 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-config-data\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.304612 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39cfebb3-d0da-4535-8c1c-81d1db536ae1" path="/var/lib/kubelet/pods/39cfebb3-d0da-4535-8c1c-81d1db536ae1/volumes" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.364001 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.364073 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-config-data\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.364146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-internal-tls-certs\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.364166 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-scripts\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.364198 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-public-tls-certs\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.364238 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfrjk\" (UniqueName: \"kubernetes.io/projected/00e7d3a7-76c8-4657-9168-5086d95c429a-kube-api-access-mfrjk\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.368707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-scripts\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.368900 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.370126 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-internal-tls-certs\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.371466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-config-data\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.372457 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e7d3a7-76c8-4657-9168-5086d95c429a-public-tls-certs\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.382042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfrjk\" (UniqueName: \"kubernetes.io/projected/00e7d3a7-76c8-4657-9168-5086d95c429a-kube-api-access-mfrjk\") pod \"aodh-0\" (UID: \"00e7d3a7-76c8-4657-9168-5086d95c429a\") " pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.519254 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 09:15:52 crc kubenswrapper[4744]: I1201 09:15:52.979048 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 09:15:53 crc kubenswrapper[4744]: I1201 09:15:53.106556 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"00e7d3a7-76c8-4657-9168-5086d95c429a","Type":"ContainerStarted","Data":"66810e3da32c04ef9d8a7c2e030b64295db3c1a840fdfc02e195bc13109c8dfa"} Dec 01 09:15:54 crc kubenswrapper[4744]: I1201 09:15:54.117202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"00e7d3a7-76c8-4657-9168-5086d95c429a","Type":"ContainerStarted","Data":"7011bdb61fd6ab3faa9736b38106753b4fd260733887a00f3f17a426b8ecfce8"} Dec 01 09:15:55 crc kubenswrapper[4744]: I1201 09:15:55.131360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"00e7d3a7-76c8-4657-9168-5086d95c429a","Type":"ContainerStarted","Data":"d6f0626ff66d8145a9cd11dee1bbb86c9f4c01684124b9aad72eef1fda7984d5"} Dec 01 09:15:56 crc kubenswrapper[4744]: I1201 09:15:56.147760 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"00e7d3a7-76c8-4657-9168-5086d95c429a","Type":"ContainerStarted","Data":"0f05c0f01744d36706f2c85d3c681716e1b107e157d2353784b25e33c2839949"} Dec 01 09:15:57 crc kubenswrapper[4744]: I1201 09:15:57.162238 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"00e7d3a7-76c8-4657-9168-5086d95c429a","Type":"ContainerStarted","Data":"ffeceb09c4851b44afdc46d6b3468d43717757a7f57322a8353b1301a4f3cddf"} Dec 01 09:15:57 crc kubenswrapper[4744]: I1201 09:15:57.187969 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.62583909 podStartE2EDuration="5.187951609s" podCreationTimestamp="2025-12-01 09:15:52 +0000 UTC" firstStartedPulling="2025-12-01 09:15:52.980741888 +0000 UTC m=+3484.969799809" lastFinishedPulling="2025-12-01 09:15:56.542854407 +0000 UTC m=+3488.531912328" observedRunningTime="2025-12-01 09:15:57.181230007 +0000 UTC m=+3489.170287918" watchObservedRunningTime="2025-12-01 09:15:57.187951609 +0000 UTC m=+3489.177009520" Dec 01 09:16:04 crc kubenswrapper[4744]: I1201 09:16:04.287740 4744 scope.go:117] "RemoveContainer" containerID="baea8f9cd6953899c97587599aab7427eb4d2c779e5ce2cfdac40736d46a606a" Dec 01 09:16:05 crc kubenswrapper[4744]: I1201 09:16:05.051467 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-f4krx"] Dec 01 09:16:05 crc kubenswrapper[4744]: I1201 09:16:05.061245 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-f4krx"] Dec 01 09:16:06 crc kubenswrapper[4744]: I1201 09:16:06.030512 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-b02d-account-create-update-vkxp4"] Dec 01 09:16:06 crc kubenswrapper[4744]: I1201 09:16:06.039359 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-b02d-account-create-update-vkxp4"] Dec 01 09:16:06 crc kubenswrapper[4744]: I1201 09:16:06.310616 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d75d787-92a2-4fe3-8783-73d3029693a9" path="/var/lib/kubelet/pods/2d75d787-92a2-4fe3-8783-73d3029693a9/volumes" Dec 01 09:16:06 crc kubenswrapper[4744]: I1201 09:16:06.311255 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c983cab-0958-4f0d-964c-1a188cad056e" path="/var/lib/kubelet/pods/6c983cab-0958-4f0d-964c-1a188cad056e/volumes" Dec 01 09:16:17 crc kubenswrapper[4744]: I1201 09:16:17.036591 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-tlszx"] Dec 01 09:16:17 crc kubenswrapper[4744]: I1201 09:16:17.045370 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-tlszx"] Dec 01 09:16:18 crc kubenswrapper[4744]: I1201 09:16:18.296068 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf79a6a4-97c1-4547-9d12-05c8d57534a8" path="/var/lib/kubelet/pods/bf79a6a4-97c1-4547-9d12-05c8d57534a8/volumes" Dec 01 09:16:18 crc kubenswrapper[4744]: I1201 09:16:18.633896 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:16:18 crc kubenswrapper[4744]: I1201 09:16:18.633969 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:16:18 crc kubenswrapper[4744]: I1201 09:16:18.634257 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 09:16:18 crc kubenswrapper[4744]: I1201 09:16:18.634956 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c19ec233f541084953b250fd9ef16a3aa9905562e8421df21f2178ae31c491d1"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:16:18 crc kubenswrapper[4744]: I1201 09:16:18.635022 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://c19ec233f541084953b250fd9ef16a3aa9905562e8421df21f2178ae31c491d1" gracePeriod=600 Dec 01 09:16:19 crc kubenswrapper[4744]: I1201 09:16:19.393441 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="c19ec233f541084953b250fd9ef16a3aa9905562e8421df21f2178ae31c491d1" exitCode=0 Dec 01 09:16:19 crc kubenswrapper[4744]: I1201 09:16:19.393514 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"c19ec233f541084953b250fd9ef16a3aa9905562e8421df21f2178ae31c491d1"} Dec 01 09:16:19 crc kubenswrapper[4744]: I1201 09:16:19.394109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e"} Dec 01 09:16:19 crc kubenswrapper[4744]: I1201 09:16:19.394136 4744 scope.go:117] "RemoveContainer" containerID="849365e5541d6b33b0f7ee5d258ca7eb4ed383498542a08cfe8fe908d747e471" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.289694 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nwfv4"] Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.293464 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.331438 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwfv4"] Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.427719 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-utilities\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.427821 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqhg6\" (UniqueName: \"kubernetes.io/projected/7ff629a7-09b5-4559-aabc-73c3237363e1-kube-api-access-cqhg6\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.429067 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-catalog-content\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.531205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-catalog-content\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.531311 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-utilities\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.531379 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqhg6\" (UniqueName: \"kubernetes.io/projected/7ff629a7-09b5-4559-aabc-73c3237363e1-kube-api-access-cqhg6\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.532187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-catalog-content\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.532400 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-utilities\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.555490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqhg6\" (UniqueName: \"kubernetes.io/projected/7ff629a7-09b5-4559-aabc-73c3237363e1-kube-api-access-cqhg6\") pod \"redhat-marketplace-nwfv4\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:25 crc kubenswrapper[4744]: I1201 09:16:25.618357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:26 crc kubenswrapper[4744]: I1201 09:16:26.127554 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwfv4"] Dec 01 09:16:26 crc kubenswrapper[4744]: I1201 09:16:26.496301 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerID="124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c" exitCode=0 Dec 01 09:16:26 crc kubenswrapper[4744]: I1201 09:16:26.496354 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwfv4" event={"ID":"7ff629a7-09b5-4559-aabc-73c3237363e1","Type":"ContainerDied","Data":"124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c"} Dec 01 09:16:26 crc kubenswrapper[4744]: I1201 09:16:26.496385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwfv4" event={"ID":"7ff629a7-09b5-4559-aabc-73c3237363e1","Type":"ContainerStarted","Data":"0de647d663dee27dedc2ceafe17e335ed259cbcf8ad751adacb866f056075441"} Dec 01 09:16:26 crc kubenswrapper[4744]: I1201 09:16:26.498710 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:16:27 crc kubenswrapper[4744]: I1201 09:16:27.527799 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwfv4" event={"ID":"7ff629a7-09b5-4559-aabc-73c3237363e1","Type":"ContainerStarted","Data":"82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6"} Dec 01 09:16:28 crc kubenswrapper[4744]: I1201 09:16:28.539149 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerID="82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6" exitCode=0 Dec 01 09:16:28 crc kubenswrapper[4744]: I1201 09:16:28.539229 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwfv4" event={"ID":"7ff629a7-09b5-4559-aabc-73c3237363e1","Type":"ContainerDied","Data":"82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6"} Dec 01 09:16:30 crc kubenswrapper[4744]: I1201 09:16:30.581222 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwfv4" event={"ID":"7ff629a7-09b5-4559-aabc-73c3237363e1","Type":"ContainerStarted","Data":"40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918"} Dec 01 09:16:35 crc kubenswrapper[4744]: I1201 09:16:35.620184 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:35 crc kubenswrapper[4744]: I1201 09:16:35.620731 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:35 crc kubenswrapper[4744]: I1201 09:16:35.689882 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:35 crc kubenswrapper[4744]: I1201 09:16:35.708345 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nwfv4" podStartSLOduration=7.433431896 podStartE2EDuration="10.708325379s" podCreationTimestamp="2025-12-01 09:16:25 +0000 UTC" firstStartedPulling="2025-12-01 09:16:26.498429131 +0000 UTC m=+3518.487487052" lastFinishedPulling="2025-12-01 09:16:29.773322614 +0000 UTC m=+3521.762380535" observedRunningTime="2025-12-01 09:16:30.600058154 +0000 UTC m=+3522.589116075" watchObservedRunningTime="2025-12-01 09:16:35.708325379 +0000 UTC m=+3527.697383290" Dec 01 09:16:36 crc kubenswrapper[4744]: I1201 09:16:36.686907 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:36 crc kubenswrapper[4744]: I1201 09:16:36.745441 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwfv4"] Dec 01 09:16:38 crc kubenswrapper[4744]: I1201 09:16:38.662465 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nwfv4" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="registry-server" containerID="cri-o://40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918" gracePeriod=2 Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.118184 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.284147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-catalog-content\") pod \"7ff629a7-09b5-4559-aabc-73c3237363e1\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.284264 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqhg6\" (UniqueName: \"kubernetes.io/projected/7ff629a7-09b5-4559-aabc-73c3237363e1-kube-api-access-cqhg6\") pod \"7ff629a7-09b5-4559-aabc-73c3237363e1\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.284374 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-utilities\") pod \"7ff629a7-09b5-4559-aabc-73c3237363e1\" (UID: \"7ff629a7-09b5-4559-aabc-73c3237363e1\") " Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.285450 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-utilities" (OuterVolumeSpecName: "utilities") pod "7ff629a7-09b5-4559-aabc-73c3237363e1" (UID: "7ff629a7-09b5-4559-aabc-73c3237363e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.299984 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff629a7-09b5-4559-aabc-73c3237363e1-kube-api-access-cqhg6" (OuterVolumeSpecName: "kube-api-access-cqhg6") pod "7ff629a7-09b5-4559-aabc-73c3237363e1" (UID: "7ff629a7-09b5-4559-aabc-73c3237363e1"). InnerVolumeSpecName "kube-api-access-cqhg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.321283 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ff629a7-09b5-4559-aabc-73c3237363e1" (UID: "7ff629a7-09b5-4559-aabc-73c3237363e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.387547 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.387574 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqhg6\" (UniqueName: \"kubernetes.io/projected/7ff629a7-09b5-4559-aabc-73c3237363e1-kube-api-access-cqhg6\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.387585 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff629a7-09b5-4559-aabc-73c3237363e1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.674831 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerID="40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918" exitCode=0 Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.674927 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwfv4" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.674941 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwfv4" event={"ID":"7ff629a7-09b5-4559-aabc-73c3237363e1","Type":"ContainerDied","Data":"40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918"} Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.675554 4744 scope.go:117] "RemoveContainer" containerID="40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.675360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwfv4" event={"ID":"7ff629a7-09b5-4559-aabc-73c3237363e1","Type":"ContainerDied","Data":"0de647d663dee27dedc2ceafe17e335ed259cbcf8ad751adacb866f056075441"} Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.723080 4744 scope.go:117] "RemoveContainer" containerID="82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.723196 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwfv4"] Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.732428 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwfv4"] Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.748901 4744 scope.go:117] "RemoveContainer" containerID="124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.791564 4744 scope.go:117] "RemoveContainer" containerID="40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918" Dec 01 09:16:39 crc kubenswrapper[4744]: E1201 09:16:39.791909 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918\": container with ID starting with 40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918 not found: ID does not exist" containerID="40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.791966 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918"} err="failed to get container status \"40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918\": rpc error: code = NotFound desc = could not find container \"40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918\": container with ID starting with 40af281584cb9df807fd20f7a91bb3b5f1be8524176e5a37c4a797b40e153918 not found: ID does not exist" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.792002 4744 scope.go:117] "RemoveContainer" containerID="82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6" Dec 01 09:16:39 crc kubenswrapper[4744]: E1201 09:16:39.792378 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6\": container with ID starting with 82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6 not found: ID does not exist" containerID="82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.792422 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6"} err="failed to get container status \"82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6\": rpc error: code = NotFound desc = could not find container \"82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6\": container with ID starting with 82950b3483d4c9811778708e66e5da3ecc330a68d64b17336a74af855014acc6 not found: ID does not exist" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.792440 4744 scope.go:117] "RemoveContainer" containerID="124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c" Dec 01 09:16:39 crc kubenswrapper[4744]: E1201 09:16:39.792623 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c\": container with ID starting with 124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c not found: ID does not exist" containerID="124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c" Dec 01 09:16:39 crc kubenswrapper[4744]: I1201 09:16:39.792649 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c"} err="failed to get container status \"124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c\": rpc error: code = NotFound desc = could not find container \"124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c\": container with ID starting with 124f161020a30365dd35776401e6f0cab7a413c4a70f23a2cd566fa71db2f76c not found: ID does not exist" Dec 01 09:16:40 crc kubenswrapper[4744]: I1201 09:16:40.298905 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" path="/var/lib/kubelet/pods/7ff629a7-09b5-4559-aabc-73c3237363e1/volumes" Dec 01 09:17:04 crc kubenswrapper[4744]: I1201 09:17:04.382577 4744 scope.go:117] "RemoveContainer" containerID="ce7d1a473b3058d85746221f40269742773a94dfe0266450ac3b299f125b1993" Dec 01 09:17:04 crc kubenswrapper[4744]: I1201 09:17:04.407616 4744 scope.go:117] "RemoveContainer" containerID="27dd7ce83dae3b277bcb9185ef4f981f9f854b19b7c8820266053217f5e5e03a" Dec 01 09:17:04 crc kubenswrapper[4744]: I1201 09:17:04.542351 4744 scope.go:117] "RemoveContainer" containerID="c9641cc77d338fe39b40598726b700dc87cb789e0818a0b286387c7538d98e1c" Dec 01 09:17:42 crc kubenswrapper[4744]: I1201 09:17:42.505331 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.028320 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.031380 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="prometheus" containerID="cri-o://49586ba69df19a1e1d1f847eefac4279a25497755f28a7723ff03db86b58080d" gracePeriod=600 Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.031483 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="config-reloader" containerID="cri-o://c779f4c0285b53f8939ed894af7c1bdab911906297fa7af4846d96b0bfbb37c2" gracePeriod=600 Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.031499 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="thanos-sidecar" containerID="cri-o://d3b67176fdf366f4501f6d424406e7857f33ba75b8ef1cf4f99a71f902db2517" gracePeriod=600 Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.362164 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerID="d3b67176fdf366f4501f6d424406e7857f33ba75b8ef1cf4f99a71f902db2517" exitCode=0 Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.362216 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerID="c779f4c0285b53f8939ed894af7c1bdab911906297fa7af4846d96b0bfbb37c2" exitCode=0 Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.362227 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerID="49586ba69df19a1e1d1f847eefac4279a25497755f28a7723ff03db86b58080d" exitCode=0 Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.362250 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerDied","Data":"d3b67176fdf366f4501f6d424406e7857f33ba75b8ef1cf4f99a71f902db2517"} Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.362280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerDied","Data":"c779f4c0285b53f8939ed894af7c1bdab911906297fa7af4846d96b0bfbb37c2"} Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.362294 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerDied","Data":"49586ba69df19a1e1d1f847eefac4279a25497755f28a7723ff03db86b58080d"} Dec 01 09:17:46 crc kubenswrapper[4744]: E1201 09:17:46.568845 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee2ace4e_75bf_4869_88d7_19d0f55a7bc3.slice/crio-conmon-c779f4c0285b53f8939ed894af7c1bdab911906297fa7af4846d96b0bfbb37c2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee2ace4e_75bf_4869_88d7_19d0f55a7bc3.slice/crio-c779f4c0285b53f8939ed894af7c1bdab911906297fa7af4846d96b0bfbb37c2.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.629079 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.734104 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.734390 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-thanos-prometheus-http-client-file\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.734570 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-rulefiles-0\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.734682 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-tls-assets\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.734805 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-secret-combined-ca-bundle\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.734892 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.735019 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config-out\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.735119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.735476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-db\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.735578 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.735654 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9c79\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-kube-api-access-q9c79\") pod \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\" (UID: \"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3\") " Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.735252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.735983 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-db" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "prometheus-metric-storage-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.737059 4744 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-db\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.737097 4744 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.741074 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.741109 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.741093 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-kube-api-access-q9c79" (OuterVolumeSpecName: "kube-api-access-q9c79") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "kube-api-access-q9c79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.741455 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.741500 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.741904 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config-out" (OuterVolumeSpecName: "config-out") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.743597 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config" (OuterVolumeSpecName: "config") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.746846 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.825667 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config" (OuterVolumeSpecName: "web-config") pod "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" (UID: "ee2ace4e-75bf-4869-88d7-19d0f55a7bc3"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839248 4744 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839280 4744 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839293 4744 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839303 4744 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839313 4744 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839323 4744 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839332 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9c79\" (UniqueName: \"kubernetes.io/projected/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-kube-api-access-q9c79\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839340 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:46 crc kubenswrapper[4744]: I1201 09:17:46.839349 4744 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.374780 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ee2ace4e-75bf-4869-88d7-19d0f55a7bc3","Type":"ContainerDied","Data":"7d61ba4b1118c979921f4f886981d6b2a271a7c66dbf09e7573ef53a79ceccbf"} Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.375181 4744 scope.go:117] "RemoveContainer" containerID="d3b67176fdf366f4501f6d424406e7857f33ba75b8ef1cf4f99a71f902db2517" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.375107 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.396265 4744 scope.go:117] "RemoveContainer" containerID="c779f4c0285b53f8939ed894af7c1bdab911906297fa7af4846d96b0bfbb37c2" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.416468 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.416810 4744 scope.go:117] "RemoveContainer" containerID="49586ba69df19a1e1d1f847eefac4279a25497755f28a7723ff03db86b58080d" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.427302 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.459711 4744 scope.go:117] "RemoveContainer" containerID="ef0ec917e4bf513772fac217a7b3b0f84f1b7f7e2d9ad2cce73654caf977aede" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.459892 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:17:47 crc kubenswrapper[4744]: E1201 09:17:47.460363 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="extract-content" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460377 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="extract-content" Dec 01 09:17:47 crc kubenswrapper[4744]: E1201 09:17:47.460396 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="prometheus" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460420 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="prometheus" Dec 01 09:17:47 crc kubenswrapper[4744]: E1201 09:17:47.460437 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="init-config-reloader" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460445 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="init-config-reloader" Dec 01 09:17:47 crc kubenswrapper[4744]: E1201 09:17:47.460468 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="registry-server" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460477 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="registry-server" Dec 01 09:17:47 crc kubenswrapper[4744]: E1201 09:17:47.460493 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="extract-utilities" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460502 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="extract-utilities" Dec 01 09:17:47 crc kubenswrapper[4744]: E1201 09:17:47.460516 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="config-reloader" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460524 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="config-reloader" Dec 01 09:17:47 crc kubenswrapper[4744]: E1201 09:17:47.460552 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="thanos-sidecar" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460560 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="thanos-sidecar" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460775 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="prometheus" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460797 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff629a7-09b5-4559-aabc-73c3237363e1" containerName="registry-server" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460815 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="thanos-sidecar" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.460828 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" containerName="config-reloader" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.463009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.464737 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.465874 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.466044 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.466176 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.466427 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fdbml" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.474322 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.486918 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.492353 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554206 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-config\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554239 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554273 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554295 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4blw\" (UniqueName: \"kubernetes.io/projected/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-kube-api-access-w4blw\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554356 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554388 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554436 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554476 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.554503 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656009 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656079 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-config\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656176 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656204 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4blw\" (UniqueName: \"kubernetes.io/projected/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-kube-api-access-w4blw\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656846 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.656946 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.657344 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.657629 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.660982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.661639 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.661850 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.662569 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.666101 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.666624 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-config\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.668157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.677546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.679562 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4blw\" (UniqueName: \"kubernetes.io/projected/8243e221-4e33-4bd8-bc4b-67f0e0a0d170-kube-api-access-w4blw\") pod \"prometheus-metric-storage-0\" (UID: \"8243e221-4e33-4bd8-bc4b-67f0e0a0d170\") " pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:47 crc kubenswrapper[4744]: I1201 09:17:47.834822 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 09:17:48 crc kubenswrapper[4744]: I1201 09:17:48.302313 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee2ace4e-75bf-4869-88d7-19d0f55a7bc3" path="/var/lib/kubelet/pods/ee2ace4e-75bf-4869-88d7-19d0f55a7bc3/volumes" Dec 01 09:17:48 crc kubenswrapper[4744]: I1201 09:17:48.539457 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 09:17:49 crc kubenswrapper[4744]: I1201 09:17:49.396680 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8243e221-4e33-4bd8-bc4b-67f0e0a0d170","Type":"ContainerStarted","Data":"2c97b1dbb9059afd52d0f9f225fbec4c817b9b7f0ec6e725c6b430e7c17a653e"} Dec 01 09:17:52 crc kubenswrapper[4744]: I1201 09:17:52.427020 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8243e221-4e33-4bd8-bc4b-67f0e0a0d170","Type":"ContainerStarted","Data":"130fb5269d4ba0c18b4f52c8ce4cd2dfabed7cf806347de526b636a51d466bed"} Dec 01 09:17:59 crc kubenswrapper[4744]: I1201 09:17:59.512915 4744 generic.go:334] "Generic (PLEG): container finished" podID="8243e221-4e33-4bd8-bc4b-67f0e0a0d170" containerID="130fb5269d4ba0c18b4f52c8ce4cd2dfabed7cf806347de526b636a51d466bed" exitCode=0 Dec 01 09:17:59 crc kubenswrapper[4744]: I1201 09:17:59.513020 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8243e221-4e33-4bd8-bc4b-67f0e0a0d170","Type":"ContainerDied","Data":"130fb5269d4ba0c18b4f52c8ce4cd2dfabed7cf806347de526b636a51d466bed"} Dec 01 09:18:00 crc kubenswrapper[4744]: I1201 09:18:00.526136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8243e221-4e33-4bd8-bc4b-67f0e0a0d170","Type":"ContainerStarted","Data":"56ece4975fa53a3fdddd0f563ec76a9c7fc6c1522fc92c92e7bb2844d54d30f5"} Dec 01 09:18:03 crc kubenswrapper[4744]: I1201 09:18:03.604006 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8243e221-4e33-4bd8-bc4b-67f0e0a0d170","Type":"ContainerStarted","Data":"aa6654aa87a75b97c92a229155e488d69c64dcf9f5c5a2c11208fb4e88c1d88d"} Dec 01 09:18:03 crc kubenswrapper[4744]: I1201 09:18:03.604782 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8243e221-4e33-4bd8-bc4b-67f0e0a0d170","Type":"ContainerStarted","Data":"0ccbecfdf8170dc636c7a526a168bd79ef781ebe67778c06a3d6903c0b438861"} Dec 01 09:18:03 crc kubenswrapper[4744]: I1201 09:18:03.650942 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.650915496 podStartE2EDuration="16.650915496s" podCreationTimestamp="2025-12-01 09:17:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:18:03.642631371 +0000 UTC m=+3615.631689302" watchObservedRunningTime="2025-12-01 09:18:03.650915496 +0000 UTC m=+3615.639973417" Dec 01 09:18:07 crc kubenswrapper[4744]: I1201 09:18:07.835739 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 09:18:17 crc kubenswrapper[4744]: I1201 09:18:17.835936 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 09:18:17 crc kubenswrapper[4744]: I1201 09:18:17.842470 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 09:18:18 crc kubenswrapper[4744]: I1201 09:18:18.634062 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:18:18 crc kubenswrapper[4744]: I1201 09:18:18.634133 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:18:18 crc kubenswrapper[4744]: I1201 09:18:18.766742 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 09:18:48 crc kubenswrapper[4744]: I1201 09:18:48.633889 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:18:48 crc kubenswrapper[4744]: I1201 09:18:48.634544 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:19:18 crc kubenswrapper[4744]: I1201 09:19:18.634821 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:19:18 crc kubenswrapper[4744]: I1201 09:19:18.635514 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:19:18 crc kubenswrapper[4744]: I1201 09:19:18.635595 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 09:19:18 crc kubenswrapper[4744]: I1201 09:19:18.636473 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:19:18 crc kubenswrapper[4744]: I1201 09:19:18.636542 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" gracePeriod=600 Dec 01 09:19:18 crc kubenswrapper[4744]: E1201 09:19:18.768842 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:19:19 crc kubenswrapper[4744]: I1201 09:19:19.333510 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" exitCode=0 Dec 01 09:19:19 crc kubenswrapper[4744]: I1201 09:19:19.333593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e"} Dec 01 09:19:19 crc kubenswrapper[4744]: I1201 09:19:19.333882 4744 scope.go:117] "RemoveContainer" containerID="c19ec233f541084953b250fd9ef16a3aa9905562e8421df21f2178ae31c491d1" Dec 01 09:19:19 crc kubenswrapper[4744]: I1201 09:19:19.334562 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:19:19 crc kubenswrapper[4744]: E1201 09:19:19.334815 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:19:31 crc kubenswrapper[4744]: I1201 09:19:31.285322 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:19:31 crc kubenswrapper[4744]: E1201 09:19:31.286609 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:19:42 crc kubenswrapper[4744]: I1201 09:19:42.285044 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:19:42 crc kubenswrapper[4744]: E1201 09:19:42.286733 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:19:45 crc kubenswrapper[4744]: I1201 09:19:45.927629 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:19:55 crc kubenswrapper[4744]: I1201 09:19:55.285543 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:19:55 crc kubenswrapper[4744]: E1201 09:19:55.286286 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.145992 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4vjqg/must-gather-lwxpd"] Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.147905 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.159184 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4vjqg"/"default-dockercfg-dfpgf" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.159376 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4vjqg"/"kube-root-ca.crt" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.161972 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4vjqg/must-gather-lwxpd"] Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.162587 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4vjqg"/"openshift-service-ca.crt" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.280609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/26a046ad-bfed-4596-af5c-5c1f8854a826-must-gather-output\") pod \"must-gather-lwxpd\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.280939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl48t\" (UniqueName: \"kubernetes.io/projected/26a046ad-bfed-4596-af5c-5c1f8854a826-kube-api-access-gl48t\") pod \"must-gather-lwxpd\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.383398 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/26a046ad-bfed-4596-af5c-5c1f8854a826-must-gather-output\") pod \"must-gather-lwxpd\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.383572 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl48t\" (UniqueName: \"kubernetes.io/projected/26a046ad-bfed-4596-af5c-5c1f8854a826-kube-api-access-gl48t\") pod \"must-gather-lwxpd\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.384083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/26a046ad-bfed-4596-af5c-5c1f8854a826-must-gather-output\") pod \"must-gather-lwxpd\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.414457 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl48t\" (UniqueName: \"kubernetes.io/projected/26a046ad-bfed-4596-af5c-5c1f8854a826-kube-api-access-gl48t\") pod \"must-gather-lwxpd\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:05 crc kubenswrapper[4744]: I1201 09:20:05.464444 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:20:06 crc kubenswrapper[4744]: I1201 09:20:06.103473 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4vjqg/must-gather-lwxpd"] Dec 01 09:20:06 crc kubenswrapper[4744]: I1201 09:20:06.848091 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" event={"ID":"26a046ad-bfed-4596-af5c-5c1f8854a826","Type":"ContainerStarted","Data":"d6e31864f20f7fef7ae711a06582648290d68b47bd1a7f7206602eb22ef23f68"} Dec 01 09:20:10 crc kubenswrapper[4744]: I1201 09:20:10.285302 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:20:10 crc kubenswrapper[4744]: E1201 09:20:10.286179 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:20:10 crc kubenswrapper[4744]: I1201 09:20:10.896818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" event={"ID":"26a046ad-bfed-4596-af5c-5c1f8854a826","Type":"ContainerStarted","Data":"28a0e804cfdeff7ede950f0a07b03984c9dbb6d8b782950fdba615c7845801ed"} Dec 01 09:20:10 crc kubenswrapper[4744]: I1201 09:20:10.897167 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" event={"ID":"26a046ad-bfed-4596-af5c-5c1f8854a826","Type":"ContainerStarted","Data":"f7d89553cc1ce987088120c6b33715e47550d0f4f40b4b11460e2f92a07470d4"} Dec 01 09:20:10 crc kubenswrapper[4744]: I1201 09:20:10.920242 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" podStartSLOduration=1.974346073 podStartE2EDuration="5.920218994s" podCreationTimestamp="2025-12-01 09:20:05 +0000 UTC" firstStartedPulling="2025-12-01 09:20:06.103127851 +0000 UTC m=+3738.092185762" lastFinishedPulling="2025-12-01 09:20:10.049000722 +0000 UTC m=+3742.038058683" observedRunningTime="2025-12-01 09:20:10.913230895 +0000 UTC m=+3742.902288836" watchObservedRunningTime="2025-12-01 09:20:10.920218994 +0000 UTC m=+3742.909276915" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.147936 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4vjqg/crc-debug-jqlsp"] Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.149772 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.214366 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv99g\" (UniqueName: \"kubernetes.io/projected/07356095-50fe-4256-a728-ace921e3692b-kube-api-access-gv99g\") pod \"crc-debug-jqlsp\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.214445 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07356095-50fe-4256-a728-ace921e3692b-host\") pod \"crc-debug-jqlsp\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.316884 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv99g\" (UniqueName: \"kubernetes.io/projected/07356095-50fe-4256-a728-ace921e3692b-kube-api-access-gv99g\") pod \"crc-debug-jqlsp\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.316982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07356095-50fe-4256-a728-ace921e3692b-host\") pod \"crc-debug-jqlsp\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.317083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07356095-50fe-4256-a728-ace921e3692b-host\") pod \"crc-debug-jqlsp\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.348053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv99g\" (UniqueName: \"kubernetes.io/projected/07356095-50fe-4256-a728-ace921e3692b-kube-api-access-gv99g\") pod \"crc-debug-jqlsp\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.466858 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:15 crc kubenswrapper[4744]: I1201 09:20:15.939808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" event={"ID":"07356095-50fe-4256-a728-ace921e3692b","Type":"ContainerStarted","Data":"e27eacb5e13d16d7ce81b40a5f8592bbc323eb59814aa5703d8d8804682d6af0"} Dec 01 09:20:22 crc kubenswrapper[4744]: I1201 09:20:22.285062 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:20:22 crc kubenswrapper[4744]: E1201 09:20:22.285955 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:20:29 crc kubenswrapper[4744]: I1201 09:20:29.065376 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" event={"ID":"07356095-50fe-4256-a728-ace921e3692b","Type":"ContainerStarted","Data":"755e0d4e87fce0416921a826de0cdfe4c2083dd65a87bcf9e522997055aff25d"} Dec 01 09:20:29 crc kubenswrapper[4744]: I1201 09:20:29.090584 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" podStartSLOduration=1.729276078 podStartE2EDuration="14.090562472s" podCreationTimestamp="2025-12-01 09:20:15 +0000 UTC" firstStartedPulling="2025-12-01 09:20:15.503191509 +0000 UTC m=+3747.492249430" lastFinishedPulling="2025-12-01 09:20:27.864477903 +0000 UTC m=+3759.853535824" observedRunningTime="2025-12-01 09:20:29.077913343 +0000 UTC m=+3761.066971274" watchObservedRunningTime="2025-12-01 09:20:29.090562472 +0000 UTC m=+3761.079620403" Dec 01 09:20:33 crc kubenswrapper[4744]: I1201 09:20:33.284836 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:20:33 crc kubenswrapper[4744]: E1201 09:20:33.285665 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:20:45 crc kubenswrapper[4744]: I1201 09:20:45.284758 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:20:45 crc kubenswrapper[4744]: E1201 09:20:45.286924 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:20:46 crc kubenswrapper[4744]: I1201 09:20:46.217113 4744 generic.go:334] "Generic (PLEG): container finished" podID="07356095-50fe-4256-a728-ace921e3692b" containerID="755e0d4e87fce0416921a826de0cdfe4c2083dd65a87bcf9e522997055aff25d" exitCode=0 Dec 01 09:20:46 crc kubenswrapper[4744]: I1201 09:20:46.217285 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" event={"ID":"07356095-50fe-4256-a728-ace921e3692b","Type":"ContainerDied","Data":"755e0d4e87fce0416921a826de0cdfe4c2083dd65a87bcf9e522997055aff25d"} Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.366679 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.399670 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4vjqg/crc-debug-jqlsp"] Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.409458 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4vjqg/crc-debug-jqlsp"] Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.450380 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv99g\" (UniqueName: \"kubernetes.io/projected/07356095-50fe-4256-a728-ace921e3692b-kube-api-access-gv99g\") pod \"07356095-50fe-4256-a728-ace921e3692b\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.450650 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07356095-50fe-4256-a728-ace921e3692b-host\") pod \"07356095-50fe-4256-a728-ace921e3692b\" (UID: \"07356095-50fe-4256-a728-ace921e3692b\") " Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.450945 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07356095-50fe-4256-a728-ace921e3692b-host" (OuterVolumeSpecName: "host") pod "07356095-50fe-4256-a728-ace921e3692b" (UID: "07356095-50fe-4256-a728-ace921e3692b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.451265 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07356095-50fe-4256-a728-ace921e3692b-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.458398 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07356095-50fe-4256-a728-ace921e3692b-kube-api-access-gv99g" (OuterVolumeSpecName: "kube-api-access-gv99g") pod "07356095-50fe-4256-a728-ace921e3692b" (UID: "07356095-50fe-4256-a728-ace921e3692b"). InnerVolumeSpecName "kube-api-access-gv99g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:20:47 crc kubenswrapper[4744]: I1201 09:20:47.553757 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv99g\" (UniqueName: \"kubernetes.io/projected/07356095-50fe-4256-a728-ace921e3692b-kube-api-access-gv99g\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.238052 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e27eacb5e13d16d7ce81b40a5f8592bbc323eb59814aa5703d8d8804682d6af0" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.238350 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-jqlsp" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.297362 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07356095-50fe-4256-a728-ace921e3692b" path="/var/lib/kubelet/pods/07356095-50fe-4256-a728-ace921e3692b/volumes" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.583547 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4vjqg/crc-debug-bzg6j"] Dec 01 09:20:48 crc kubenswrapper[4744]: E1201 09:20:48.584093 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07356095-50fe-4256-a728-ace921e3692b" containerName="container-00" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.584116 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="07356095-50fe-4256-a728-ace921e3692b" containerName="container-00" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.584860 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="07356095-50fe-4256-a728-ace921e3692b" containerName="container-00" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.586526 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.675554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p9b7\" (UniqueName: \"kubernetes.io/projected/b1ce6f23-8852-4249-822c-8cf048f47ded-kube-api-access-5p9b7\") pod \"crc-debug-bzg6j\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.675643 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1ce6f23-8852-4249-822c-8cf048f47ded-host\") pod \"crc-debug-bzg6j\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.778237 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p9b7\" (UniqueName: \"kubernetes.io/projected/b1ce6f23-8852-4249-822c-8cf048f47ded-kube-api-access-5p9b7\") pod \"crc-debug-bzg6j\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.778388 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1ce6f23-8852-4249-822c-8cf048f47ded-host\") pod \"crc-debug-bzg6j\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.778610 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1ce6f23-8852-4249-822c-8cf048f47ded-host\") pod \"crc-debug-bzg6j\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.801107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p9b7\" (UniqueName: \"kubernetes.io/projected/b1ce6f23-8852-4249-822c-8cf048f47ded-kube-api-access-5p9b7\") pod \"crc-debug-bzg6j\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:48 crc kubenswrapper[4744]: I1201 09:20:48.904449 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:49 crc kubenswrapper[4744]: I1201 09:20:49.247252 4744 generic.go:334] "Generic (PLEG): container finished" podID="b1ce6f23-8852-4249-822c-8cf048f47ded" containerID="4a1bfbcb03c2e2586dda98cb13ea43727180437b86b91dae6b8cecb6d35fef2f" exitCode=1 Dec 01 09:20:49 crc kubenswrapper[4744]: I1201 09:20:49.247368 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" event={"ID":"b1ce6f23-8852-4249-822c-8cf048f47ded","Type":"ContainerDied","Data":"4a1bfbcb03c2e2586dda98cb13ea43727180437b86b91dae6b8cecb6d35fef2f"} Dec 01 09:20:49 crc kubenswrapper[4744]: I1201 09:20:49.247600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" event={"ID":"b1ce6f23-8852-4249-822c-8cf048f47ded","Type":"ContainerStarted","Data":"b6c9f2407450e95c3699a194e50c15851973548f1391ef208afb1a0897dccc50"} Dec 01 09:20:49 crc kubenswrapper[4744]: I1201 09:20:49.288243 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4vjqg/crc-debug-bzg6j"] Dec 01 09:20:49 crc kubenswrapper[4744]: I1201 09:20:49.301054 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4vjqg/crc-debug-bzg6j"] Dec 01 09:20:50 crc kubenswrapper[4744]: I1201 09:20:50.780011 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:50 crc kubenswrapper[4744]: I1201 09:20:50.815726 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p9b7\" (UniqueName: \"kubernetes.io/projected/b1ce6f23-8852-4249-822c-8cf048f47ded-kube-api-access-5p9b7\") pod \"b1ce6f23-8852-4249-822c-8cf048f47ded\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " Dec 01 09:20:50 crc kubenswrapper[4744]: I1201 09:20:50.815990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1ce6f23-8852-4249-822c-8cf048f47ded-host\") pod \"b1ce6f23-8852-4249-822c-8cf048f47ded\" (UID: \"b1ce6f23-8852-4249-822c-8cf048f47ded\") " Dec 01 09:20:50 crc kubenswrapper[4744]: I1201 09:20:50.816132 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1ce6f23-8852-4249-822c-8cf048f47ded-host" (OuterVolumeSpecName: "host") pod "b1ce6f23-8852-4249-822c-8cf048f47ded" (UID: "b1ce6f23-8852-4249-822c-8cf048f47ded"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:20:50 crc kubenswrapper[4744]: I1201 09:20:50.816931 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1ce6f23-8852-4249-822c-8cf048f47ded-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:50 crc kubenswrapper[4744]: I1201 09:20:50.876603 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1ce6f23-8852-4249-822c-8cf048f47ded-kube-api-access-5p9b7" (OuterVolumeSpecName: "kube-api-access-5p9b7") pod "b1ce6f23-8852-4249-822c-8cf048f47ded" (UID: "b1ce6f23-8852-4249-822c-8cf048f47ded"). InnerVolumeSpecName "kube-api-access-5p9b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:20:50 crc kubenswrapper[4744]: I1201 09:20:50.918869 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p9b7\" (UniqueName: \"kubernetes.io/projected/b1ce6f23-8852-4249-822c-8cf048f47ded-kube-api-access-5p9b7\") on node \"crc\" DevicePath \"\"" Dec 01 09:20:51 crc kubenswrapper[4744]: I1201 09:20:51.266636 4744 scope.go:117] "RemoveContainer" containerID="4a1bfbcb03c2e2586dda98cb13ea43727180437b86b91dae6b8cecb6d35fef2f" Dec 01 09:20:51 crc kubenswrapper[4744]: I1201 09:20:51.266821 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/crc-debug-bzg6j" Dec 01 09:20:51 crc kubenswrapper[4744]: E1201 09:20:51.505350 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1ce6f23_8852_4249_822c_8cf048f47ded.slice/crio-b6c9f2407450e95c3699a194e50c15851973548f1391ef208afb1a0897dccc50\": RecentStats: unable to find data in memory cache]" Dec 01 09:20:52 crc kubenswrapper[4744]: I1201 09:20:52.298260 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1ce6f23-8852-4249-822c-8cf048f47ded" path="/var/lib/kubelet/pods/b1ce6f23-8852-4249-822c-8cf048f47ded/volumes" Dec 01 09:20:57 crc kubenswrapper[4744]: I1201 09:20:57.285623 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:20:57 crc kubenswrapper[4744]: E1201 09:20:57.286339 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:21:08 crc kubenswrapper[4744]: I1201 09:21:08.296887 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:21:08 crc kubenswrapper[4744]: E1201 09:21:08.298003 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:21:21 crc kubenswrapper[4744]: I1201 09:21:21.285855 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:21:21 crc kubenswrapper[4744]: E1201 09:21:21.286502 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:21:32 crc kubenswrapper[4744]: I1201 09:21:32.285889 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:21:32 crc kubenswrapper[4744]: E1201 09:21:32.287053 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:21:38 crc kubenswrapper[4744]: I1201 09:21:38.753587 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/init-config-reloader/0.log" Dec 01 09:21:39 crc kubenswrapper[4744]: I1201 09:21:39.072213 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/init-config-reloader/0.log" Dec 01 09:21:39 crc kubenswrapper[4744]: I1201 09:21:39.131882 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/alertmanager/0.log" Dec 01 09:21:39 crc kubenswrapper[4744]: I1201 09:21:39.179765 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/config-reloader/0.log" Dec 01 09:21:39 crc kubenswrapper[4744]: I1201 09:21:39.297558 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-api/0.log" Dec 01 09:21:39 crc kubenswrapper[4744]: I1201 09:21:39.345947 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-evaluator/0.log" Dec 01 09:21:39 crc kubenswrapper[4744]: I1201 09:21:39.389087 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-listener/0.log" Dec 01 09:21:39 crc kubenswrapper[4744]: I1201 09:21:39.982980 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b74d8fd6-9hczb_c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d/barbican-api/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:39.991713 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b74d8fd6-9hczb_c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d/barbican-api-log/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.046374 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-notifier/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.212379 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f8b9684d-5kjvj_02654eb7-3943-40d1-9145-04bcd38b8344/barbican-keystone-listener/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.277836 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f8b9684d-5kjvj_02654eb7-3943-40d1-9145-04bcd38b8344/barbican-keystone-listener-log/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.418646 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b8696b95c-fb5g6_a23750f1-c129-46b5-9c30-c6fa8ee54320/barbican-worker/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.521630 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b8696b95c-fb5g6_a23750f1-c129-46b5-9c30-c6fa8ee54320/barbican-worker-log/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.642033 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c_0546ddaa-13e4-481a-b1f1-a2fe6aa505f3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.827902 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/proxy-httpd/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.845589 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/ceilometer-central-agent/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.847536 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/ceilometer-notification-agent/0.log" Dec 01 09:21:40 crc kubenswrapper[4744]: I1201 09:21:40.925625 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/sg-core/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.027179 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_08db63f5-2255-45a5-b2c2-b5baaed9fbca/cinder-api-log/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.128084 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_08db63f5-2255-45a5-b2c2-b5baaed9fbca/cinder-api/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.269618 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23516ec1-d13c-4dfd-b20f-207ba2bb0d26/cinder-scheduler/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.421914 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23516ec1-d13c-4dfd-b20f-207ba2bb0d26/probe/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.519451 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k4766_1f3aaa29-4bf7-4f79-aae9-7e90d131703d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.603504 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f_e325084e-1abd-4b89-99f7-ec9e4f6f55ef/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.707353 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-5nlp9_dee5122c-f685-44a1-acd0-9fcfe3968bb7/init/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.882806 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-5nlp9_dee5122c-f685-44a1-acd0-9fcfe3968bb7/init/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.925445 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-5nlp9_dee5122c-f685-44a1-acd0-9fcfe3968bb7/dnsmasq-dns/0.log" Dec 01 09:21:41 crc kubenswrapper[4744]: I1201 09:21:41.950690 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl_2b675722-4753-4de2-8bd8-aa37599eb53e/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.087001 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a52e8d56-63ad-466d-a7e9-57c75c346a1c/glance-httpd/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.119068 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a52e8d56-63ad-466d-a7e9-57c75c346a1c/glance-log/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.160849 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7b5c1409-3b79-4680-beb3-14ffcf0fb055/glance-httpd/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.216320 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7b5c1409-3b79-4680-beb3-14ffcf0fb055/glance-log/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.717653 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5c9d7d48-5qkmn_a60bb987-1492-435c-9f53-17c3820ce1a0/heat-api/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.797361 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68d7bb6659-xxvcz_6abf6646-56a2-4700-9a37-9474036b8865/heat-cfnapi/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.900113 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5bf9f8b584-7k2r7_11429dc7-0531-4503-b66e-f6ecef318ec0/heat-engine/0.log" Dec 01 09:21:42 crc kubenswrapper[4744]: I1201 09:21:42.954067 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7_cbfdfe61-094b-48f5-a450-0fcac89b8e8a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.042381 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fttwk_b2c369ea-0c26-424a-af22-996d64ecf25d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.253938 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-77497cdd8f-s4w2q_d2a839b7-c745-45e3-9984-3f403dbf6fce/keystone-api/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.312756 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29409661-lht94_ee0b29aa-22fe-412e-ad54-43db39412cec/keystone-cron/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.371816 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_13bbf1b1-4680-4390-99ed-e0a54dd6d6fe/kube-state-metrics/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.517798 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm_b9ea1f1e-e881-4097-bd64-59ad7d48d552/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.748946 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-685cbf69b9-jrc9r_09924952-c000-422a-b471-e738246ae161/neutron-api/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.800697 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-685cbf69b9-jrc9r_09924952-c000-422a-b471-e738246ae161/neutron-httpd/0.log" Dec 01 09:21:43 crc kubenswrapper[4744]: I1201 09:21:43.910461 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp_ede60eff-c329-466b-a25d-7a309d424bf3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:44 crc kubenswrapper[4744]: I1201 09:21:44.183100 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ec302fdc-d505-4f9c-b403-14643c501988/nova-api-log/0.log" Dec 01 09:21:44 crc kubenswrapper[4744]: I1201 09:21:44.473698 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d4ab1910-715b-4982-9903-453336b14b78/nova-cell0-conductor-conductor/0.log" Dec 01 09:21:44 crc kubenswrapper[4744]: I1201 09:21:44.498271 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ec302fdc-d505-4f9c-b403-14643c501988/nova-api-api/0.log" Dec 01 09:21:44 crc kubenswrapper[4744]: I1201 09:21:44.521604 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0ae053c1-f41c-4bbb-bae8-66e1d3d891b7/nova-cell1-conductor-conductor/0.log" Dec 01 09:21:44 crc kubenswrapper[4744]: I1201 09:21:44.776647 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c9f8476b-a1a9-42b5-ba40-95be5413d23b/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 09:21:44 crc kubenswrapper[4744]: I1201 09:21:44.869202 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8rpw6_da76f156-fbfc-46fe-b11d-a71a0bfd280d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:45 crc kubenswrapper[4744]: I1201 09:21:45.099227 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7bfdbce6-2016-4da0-8d18-300bbc3e92fb/nova-metadata-log/0.log" Dec 01 09:21:45 crc kubenswrapper[4744]: I1201 09:21:45.232110 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_97f2638c-75e4-4b25-be41-16e6b261c1b3/nova-scheduler-scheduler/0.log" Dec 01 09:21:45 crc kubenswrapper[4744]: I1201 09:21:45.360104 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c31db90c-376f-45b5-9d13-73bcf4962135/mysql-bootstrap/0.log" Dec 01 09:21:45 crc kubenswrapper[4744]: I1201 09:21:45.662212 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c31db90c-376f-45b5-9d13-73bcf4962135/mysql-bootstrap/0.log" Dec 01 09:21:45 crc kubenswrapper[4744]: I1201 09:21:45.672269 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c31db90c-376f-45b5-9d13-73bcf4962135/galera/0.log" Dec 01 09:21:45 crc kubenswrapper[4744]: I1201 09:21:45.855535 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0b711f28-53bd-46e8-9a86-44be43bd78d3/mysql-bootstrap/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.049694 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0b711f28-53bd-46e8-9a86-44be43bd78d3/galera/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.057509 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0b711f28-53bd-46e8-9a86-44be43bd78d3/mysql-bootstrap/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.375875 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7bfdbce6-2016-4da0-8d18-300bbc3e92fb/nova-metadata-metadata/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.412152 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_55446f52-3cf3-4ac2-9ad1-7f35c5508697/openstackclient/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.465534 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4nncr_efccc7e5-6075-43fb-b36e-15c4734e8f97/openstack-network-exporter/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.561098 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovsdb-server-init/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.849456 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovsdb-server-init/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.851682 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovs-vswitchd/0.log" Dec 01 09:21:46 crc kubenswrapper[4744]: I1201 09:21:46.947567 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovsdb-server/0.log" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.011444 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pgp2f_ac7e9e72-1bca-42aa-a062-e9743651f2aa/ovn-controller/0.log" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.153366 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7v6p7_25afa440-2166-4a7f-a6b9-8d4dc2c7087c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.210064 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_551c38c7-428c-427a-b9b3-1322157a08df/openstack-network-exporter/0.log" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.285766 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:21:47 crc kubenswrapper[4744]: E1201 09:21:47.285966 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.330353 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_551c38c7-428c-427a-b9b3-1322157a08df/ovn-northd/0.log" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.891944 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ebb51f-ba73-417a-b6a8-ff31ff555108/openstack-network-exporter/0.log" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.933921 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ebb51f-ba73-417a-b6a8-ff31ff555108/ovsdbserver-nb/0.log" Dec 01 09:21:47 crc kubenswrapper[4744]: I1201 09:21:47.944862 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_916bb441-60a0-4452-b2cf-5b98fb9aea9f/openstack-network-exporter/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.087074 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_916bb441-60a0-4452-b2cf-5b98fb9aea9f/ovsdbserver-sb/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.262259 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86964bb7b8-4bgsm_b7d2acd7-91fd-4cc6-a9de-f7a3fb905192/placement-api/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.272262 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86964bb7b8-4bgsm_b7d2acd7-91fd-4cc6-a9de-f7a3fb905192/placement-log/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.403125 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/init-config-reloader/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.573773 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/init-config-reloader/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.610678 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/prometheus/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.641912 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/config-reloader/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.662059 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/thanos-sidecar/0.log" Dec 01 09:21:48 crc kubenswrapper[4744]: I1201 09:21:48.859329 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e4a93bac-aabe-4109-950f-14cb6ac634dd/setup-container/0.log" Dec 01 09:21:49 crc kubenswrapper[4744]: I1201 09:21:49.570890 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e4a93bac-aabe-4109-950f-14cb6ac634dd/setup-container/0.log" Dec 01 09:21:49 crc kubenswrapper[4744]: I1201 09:21:49.629198 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_93e40ea3-265b-44a9-ada5-5b583201148d/setup-container/0.log" Dec 01 09:21:49 crc kubenswrapper[4744]: I1201 09:21:49.645981 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e4a93bac-aabe-4109-950f-14cb6ac634dd/rabbitmq/0.log" Dec 01 09:21:49 crc kubenswrapper[4744]: I1201 09:21:49.857269 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_93e40ea3-265b-44a9-ada5-5b583201148d/setup-container/0.log" Dec 01 09:21:49 crc kubenswrapper[4744]: I1201 09:21:49.955688 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz_de4a3818-81e1-4735-b58d-6d1bb9df5ee2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:50 crc kubenswrapper[4744]: I1201 09:21:50.098090 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-mn867_e59297bb-e715-4a12-b214-70d9aa5c4757/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:50 crc kubenswrapper[4744]: I1201 09:21:50.566622 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz_75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:50 crc kubenswrapper[4744]: I1201 09:21:50.620908 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-25vvz_dbf61995-97c3-420f-a591-5867f9192102/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:50 crc kubenswrapper[4744]: I1201 09:21:50.682530 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_93e40ea3-265b-44a9-ada5-5b583201148d/rabbitmq/0.log" Dec 01 09:21:50 crc kubenswrapper[4744]: I1201 09:21:50.834290 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-g5snq_1e08d86e-f21d-42a3-8d93-be22013b97ac/ssh-known-hosts-edpm-deployment/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.050590 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696c89c747-77nz8_2e5f2594-2a7b-4c08-ac37-3e258f99fbb4/proxy-server/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.074309 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-l9b2z_4f171eaa-706a-47ae-be94-1d1cbd636df1/swift-ring-rebalance/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.107178 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696c89c747-77nz8_2e5f2594-2a7b-4c08-ac37-3e258f99fbb4/proxy-httpd/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.313728 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-reaper/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.354954 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-auditor/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.481067 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-replicator/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.516489 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-server/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.626036 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-auditor/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.691123 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-replicator/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.744048 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-server/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.810177 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-updater/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.901495 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-auditor/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.907308 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-expirer/0.log" Dec 01 09:21:51 crc kubenswrapper[4744]: I1201 09:21:51.986133 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-replicator/0.log" Dec 01 09:21:52 crc kubenswrapper[4744]: I1201 09:21:52.088323 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-server/0.log" Dec 01 09:21:52 crc kubenswrapper[4744]: I1201 09:21:52.114565 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/rsync/0.log" Dec 01 09:21:52 crc kubenswrapper[4744]: I1201 09:21:52.191811 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-updater/0.log" Dec 01 09:21:52 crc kubenswrapper[4744]: I1201 09:21:52.231716 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/swift-recon-cron/0.log" Dec 01 09:21:52 crc kubenswrapper[4744]: I1201 09:21:52.359132 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s24rv_ffbaa140-8ee6-4639-9524-96460ffe5a49/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:52 crc kubenswrapper[4744]: I1201 09:21:52.451118 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt_25cf1554-6aff-4169-ae0e-a6b6979ad45e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.583678 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qm5jf"] Dec 01 09:21:54 crc kubenswrapper[4744]: E1201 09:21:54.584691 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ce6f23-8852-4249-822c-8cf048f47ded" containerName="container-00" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.584706 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ce6f23-8852-4249-822c-8cf048f47ded" containerName="container-00" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.584921 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1ce6f23-8852-4249-822c-8cf048f47ded" containerName="container-00" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.588210 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.611431 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qm5jf"] Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.762485 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-utilities\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.762608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqbd8\" (UniqueName: \"kubernetes.io/projected/e119ba92-fa29-402c-a60e-fd7027d681bc-kube-api-access-sqbd8\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.762636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-catalog-content\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.865543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-utilities\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.865676 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqbd8\" (UniqueName: \"kubernetes.io/projected/e119ba92-fa29-402c-a60e-fd7027d681bc-kube-api-access-sqbd8\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.865702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-catalog-content\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.866206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-catalog-content\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.866496 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-utilities\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.883752 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqbd8\" (UniqueName: \"kubernetes.io/projected/e119ba92-fa29-402c-a60e-fd7027d681bc-kube-api-access-sqbd8\") pod \"community-operators-qm5jf\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:54 crc kubenswrapper[4744]: I1201 09:21:54.923347 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:21:55 crc kubenswrapper[4744]: I1201 09:21:55.562518 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qm5jf"] Dec 01 09:21:56 crc kubenswrapper[4744]: I1201 09:21:56.415568 4744 generic.go:334] "Generic (PLEG): container finished" podID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerID="1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d" exitCode=0 Dec 01 09:21:56 crc kubenswrapper[4744]: I1201 09:21:56.415621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qm5jf" event={"ID":"e119ba92-fa29-402c-a60e-fd7027d681bc","Type":"ContainerDied","Data":"1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d"} Dec 01 09:21:56 crc kubenswrapper[4744]: I1201 09:21:56.415899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qm5jf" event={"ID":"e119ba92-fa29-402c-a60e-fd7027d681bc","Type":"ContainerStarted","Data":"b3eb5cbf43451e2a863e29722d4f7edf42e1a6291bd6fc2cf505f4c6ca453fdf"} Dec 01 09:21:56 crc kubenswrapper[4744]: I1201 09:21:56.423893 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:21:58 crc kubenswrapper[4744]: I1201 09:21:58.436834 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qm5jf" event={"ID":"e119ba92-fa29-402c-a60e-fd7027d681bc","Type":"ContainerStarted","Data":"2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2"} Dec 01 09:21:59 crc kubenswrapper[4744]: I1201 09:21:59.286173 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:21:59 crc kubenswrapper[4744]: E1201 09:21:59.286690 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:21:59 crc kubenswrapper[4744]: I1201 09:21:59.456964 4744 generic.go:334] "Generic (PLEG): container finished" podID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerID="2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2" exitCode=0 Dec 01 09:21:59 crc kubenswrapper[4744]: I1201 09:21:59.457012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qm5jf" event={"ID":"e119ba92-fa29-402c-a60e-fd7027d681bc","Type":"ContainerDied","Data":"2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2"} Dec 01 09:22:00 crc kubenswrapper[4744]: I1201 09:22:00.013989 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0153c296-2e70-43fa-90ae-bb007406d2c0/memcached/0.log" Dec 01 09:22:00 crc kubenswrapper[4744]: I1201 09:22:00.467619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qm5jf" event={"ID":"e119ba92-fa29-402c-a60e-fd7027d681bc","Type":"ContainerStarted","Data":"a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be"} Dec 01 09:22:00 crc kubenswrapper[4744]: I1201 09:22:00.496273 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qm5jf" podStartSLOduration=2.935553933 podStartE2EDuration="6.496251828s" podCreationTimestamp="2025-12-01 09:21:54 +0000 UTC" firstStartedPulling="2025-12-01 09:21:56.423596374 +0000 UTC m=+3848.412654295" lastFinishedPulling="2025-12-01 09:21:59.984294269 +0000 UTC m=+3851.973352190" observedRunningTime="2025-12-01 09:22:00.486323407 +0000 UTC m=+3852.475381328" watchObservedRunningTime="2025-12-01 09:22:00.496251828 +0000 UTC m=+3852.485309749" Dec 01 09:22:04 crc kubenswrapper[4744]: I1201 09:22:04.924052 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:22:04 crc kubenswrapper[4744]: I1201 09:22:04.925276 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:22:05 crc kubenswrapper[4744]: I1201 09:22:05.013099 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:22:05 crc kubenswrapper[4744]: I1201 09:22:05.571003 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:22:05 crc kubenswrapper[4744]: I1201 09:22:05.629474 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qm5jf"] Dec 01 09:22:07 crc kubenswrapper[4744]: I1201 09:22:07.533786 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qm5jf" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="registry-server" containerID="cri-o://a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be" gracePeriod=2 Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.071286 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.163341 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-catalog-content\") pod \"e119ba92-fa29-402c-a60e-fd7027d681bc\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.163527 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-utilities\") pod \"e119ba92-fa29-402c-a60e-fd7027d681bc\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.163595 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqbd8\" (UniqueName: \"kubernetes.io/projected/e119ba92-fa29-402c-a60e-fd7027d681bc-kube-api-access-sqbd8\") pod \"e119ba92-fa29-402c-a60e-fd7027d681bc\" (UID: \"e119ba92-fa29-402c-a60e-fd7027d681bc\") " Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.164879 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-utilities" (OuterVolumeSpecName: "utilities") pod "e119ba92-fa29-402c-a60e-fd7027d681bc" (UID: "e119ba92-fa29-402c-a60e-fd7027d681bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.165680 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.191778 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e119ba92-fa29-402c-a60e-fd7027d681bc-kube-api-access-sqbd8" (OuterVolumeSpecName: "kube-api-access-sqbd8") pod "e119ba92-fa29-402c-a60e-fd7027d681bc" (UID: "e119ba92-fa29-402c-a60e-fd7027d681bc"). InnerVolumeSpecName "kube-api-access-sqbd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.227484 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e119ba92-fa29-402c-a60e-fd7027d681bc" (UID: "e119ba92-fa29-402c-a60e-fd7027d681bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.267760 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e119ba92-fa29-402c-a60e-fd7027d681bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.267790 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqbd8\" (UniqueName: \"kubernetes.io/projected/e119ba92-fa29-402c-a60e-fd7027d681bc-kube-api-access-sqbd8\") on node \"crc\" DevicePath \"\"" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.544564 4744 generic.go:334] "Generic (PLEG): container finished" podID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerID="a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be" exitCode=0 Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.544605 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qm5jf" event={"ID":"e119ba92-fa29-402c-a60e-fd7027d681bc","Type":"ContainerDied","Data":"a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be"} Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.544630 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qm5jf" event={"ID":"e119ba92-fa29-402c-a60e-fd7027d681bc","Type":"ContainerDied","Data":"b3eb5cbf43451e2a863e29722d4f7edf42e1a6291bd6fc2cf505f4c6ca453fdf"} Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.544652 4744 scope.go:117] "RemoveContainer" containerID="a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.545678 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qm5jf" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.567200 4744 scope.go:117] "RemoveContainer" containerID="2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.575283 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qm5jf"] Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.585894 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qm5jf"] Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.593306 4744 scope.go:117] "RemoveContainer" containerID="1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.660141 4744 scope.go:117] "RemoveContainer" containerID="a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be" Dec 01 09:22:08 crc kubenswrapper[4744]: E1201 09:22:08.660604 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be\": container with ID starting with a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be not found: ID does not exist" containerID="a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.660642 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be"} err="failed to get container status \"a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be\": rpc error: code = NotFound desc = could not find container \"a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be\": container with ID starting with a75443ac885adea7d0746527ee59c9888f6f457a47f68bf40b715f18ed8422be not found: ID does not exist" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.660670 4744 scope.go:117] "RemoveContainer" containerID="2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2" Dec 01 09:22:08 crc kubenswrapper[4744]: E1201 09:22:08.660927 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2\": container with ID starting with 2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2 not found: ID does not exist" containerID="2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.660956 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2"} err="failed to get container status \"2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2\": rpc error: code = NotFound desc = could not find container \"2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2\": container with ID starting with 2a032b152c769680cf060da59fe77c60ef56ff8ec631ddabe8e97c8d682c76b2 not found: ID does not exist" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.660974 4744 scope.go:117] "RemoveContainer" containerID="1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d" Dec 01 09:22:08 crc kubenswrapper[4744]: E1201 09:22:08.661207 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d\": container with ID starting with 1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d not found: ID does not exist" containerID="1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d" Dec 01 09:22:08 crc kubenswrapper[4744]: I1201 09:22:08.661246 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d"} err="failed to get container status \"1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d\": rpc error: code = NotFound desc = could not find container \"1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d\": container with ID starting with 1036f508325c7e182b705cd9f29c0da4b84b269be298f950f438199c885b109d not found: ID does not exist" Dec 01 09:22:10 crc kubenswrapper[4744]: I1201 09:22:10.295864 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" path="/var/lib/kubelet/pods/e119ba92-fa29-402c-a60e-fd7027d681bc/volumes" Dec 01 09:22:11 crc kubenswrapper[4744]: I1201 09:22:11.285629 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:22:11 crc kubenswrapper[4744]: E1201 09:22:11.286075 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.165194 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/util/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.288163 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:22:22 crc kubenswrapper[4744]: E1201 09:22:22.288435 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.414683 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/util/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.417492 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/pull/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.424044 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/pull/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.618044 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/pull/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.621480 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/util/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.651872 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/extract/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.811686 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-wf77m_7bc8fb90-3b43-43f8-8df4-128c639ad76a/kube-rbac-proxy/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.900909 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-wf77m_7bc8fb90-3b43-43f8-8df4-128c639ad76a/manager/0.log" Dec 01 09:22:22 crc kubenswrapper[4744]: I1201 09:22:22.947097 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-lkltp_d9f50601-c7cd-4db6-82dd-9581a5dc8a10/kube-rbac-proxy/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.072914 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-lkltp_d9f50601-c7cd-4db6-82dd-9581a5dc8a10/manager/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.161610 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-8zfjd_2fbe6884-845e-4c7d-aff5-0deb483f4d19/kube-rbac-proxy/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.193762 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-8zfjd_2fbe6884-845e-4c7d-aff5-0deb483f4d19/manager/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.366394 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-kvcrb_f5fa46a8-32d1-4eda-89ec-04bdfd35ea78/kube-rbac-proxy/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.520785 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-kvcrb_f5fa46a8-32d1-4eda-89ec-04bdfd35ea78/manager/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.577039 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-7z2ml_badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a/kube-rbac-proxy/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.702893 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-7z2ml_badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a/manager/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.743056 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pv6f6_5f4c0ab8-4dee-4c39-b630-061580e95592/kube-rbac-proxy/0.log" Dec 01 09:22:23 crc kubenswrapper[4744]: I1201 09:22:23.940738 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-dgxbb_ceb61e3f-5534-46ad-ba39-1fe0b07530f7/kube-rbac-proxy/0.log" Dec 01 09:22:24 crc kubenswrapper[4744]: I1201 09:22:24.052233 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pv6f6_5f4c0ab8-4dee-4c39-b630-061580e95592/manager/0.log" Dec 01 09:22:24 crc kubenswrapper[4744]: I1201 09:22:24.236855 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-dgxbb_ceb61e3f-5534-46ad-ba39-1fe0b07530f7/manager/0.log" Dec 01 09:22:24 crc kubenswrapper[4744]: I1201 09:22:24.323596 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-gg7s7_214d7853-ed12-4624-82cd-6b279a827a13/kube-rbac-proxy/0.log" Dec 01 09:22:24 crc kubenswrapper[4744]: I1201 09:22:24.699585 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-gg7s7_214d7853-ed12-4624-82cd-6b279a827a13/manager/0.log" Dec 01 09:22:24 crc kubenswrapper[4744]: I1201 09:22:24.758667 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-7fn8v_8b89e5df-c38d-4465-aa74-b7c67e268a0d/kube-rbac-proxy/0.log" Dec 01 09:22:24 crc kubenswrapper[4744]: I1201 09:22:24.837894 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-7fn8v_8b89e5df-c38d-4465-aa74-b7c67e268a0d/manager/0.log" Dec 01 09:22:24 crc kubenswrapper[4744]: I1201 09:22:24.974608 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-tlszh_4cf33df0-316f-4acf-b105-8dc967fae633/kube-rbac-proxy/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.008749 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-tlszh_4cf33df0-316f-4acf-b105-8dc967fae633/manager/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.173499 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-wqvcv_f68be396-2f8f-4c90-99d3-bbe12b7d6c8e/kube-rbac-proxy/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.184938 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-wqvcv_f68be396-2f8f-4c90-99d3-bbe12b7d6c8e/manager/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.312036 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rbt96_b42166df-9137-4ddb-b9b0-33515fcab298/kube-rbac-proxy/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.376687 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rbt96_b42166df-9137-4ddb-b9b0-33515fcab298/manager/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.484559 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-q9dxh_5e8a8cd5-653f-4ef0-97c6-36b1a045f59a/kube-rbac-proxy/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.585842 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-q9dxh_5e8a8cd5-653f-4ef0-97c6-36b1a045f59a/manager/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.711860 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-htgsw_c0f26427-9c16-40b5-ad62-5a7161cba83c/manager/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.718629 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-htgsw_c0f26427-9c16-40b5-ad62-5a7161cba83c/kube-rbac-proxy/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.836003 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4khmps_83c42810-1f8a-4616-9544-db48c6020c85/kube-rbac-proxy/0.log" Dec 01 09:22:25 crc kubenswrapper[4744]: I1201 09:22:25.904291 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4khmps_83c42810-1f8a-4616-9544-db48c6020c85/manager/0.log" Dec 01 09:22:26 crc kubenswrapper[4744]: I1201 09:22:26.617396 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-q498w_34cb91f3-fee7-4916-be15-b257cab23850/registry-server/0.log" Dec 01 09:22:26 crc kubenswrapper[4744]: I1201 09:22:26.824665 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d866d68c9-2fxjt_3df0fc73-2812-42d2-9c7f-e1f33591e8a1/operator/0.log" Dec 01 09:22:26 crc kubenswrapper[4744]: I1201 09:22:26.866360 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-fjx5c_102b5fb4-119f-41ba-845e-1b691b08eb57/kube-rbac-proxy/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.008695 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-fjx5c_102b5fb4-119f-41ba-845e-1b691b08eb57/manager/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.061752 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-848rm_cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5/kube-rbac-proxy/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.092233 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-848rm_cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5/manager/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.343466 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-ntg6w_b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad/kube-rbac-proxy/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.370939 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-ksvz2_c0a2f26e-7dc2-444f-aa8a-75e122512c98/operator/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.541885 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-ntg6w_b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad/manager/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.600729 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/kube-rbac-proxy/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.846196 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-bxnxg_d602d407-4c7d-4057-a87b-b7a6fb426cfe/kube-rbac-proxy/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.929924 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-bxnxg_d602d407-4c7d-4057-a87b-b7a6fb426cfe/manager/0.log" Dec 01 09:22:27 crc kubenswrapper[4744]: I1201 09:22:27.944849 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:22:28 crc kubenswrapper[4744]: I1201 09:22:28.011477 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-76f5b5884f-78kk6_717a3535-87a4-4d0a-b76d-a540a0cefa90/manager/0.log" Dec 01 09:22:28 crc kubenswrapper[4744]: I1201 09:22:28.073242 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qrbr4_3eb6d34f-5dfb-4eb6-8d81-a7f11406531f/kube-rbac-proxy/0.log" Dec 01 09:22:28 crc kubenswrapper[4744]: I1201 09:22:28.155533 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qrbr4_3eb6d34f-5dfb-4eb6-8d81-a7f11406531f/manager/0.log" Dec 01 09:22:33 crc kubenswrapper[4744]: I1201 09:22:33.285737 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:22:33 crc kubenswrapper[4744]: E1201 09:22:33.286696 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:22:45 crc kubenswrapper[4744]: I1201 09:22:45.285577 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:22:45 crc kubenswrapper[4744]: E1201 09:22:45.286226 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:22:48 crc kubenswrapper[4744]: I1201 09:22:48.512543 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dmlw7_075bdaa2-f2c7-4c5a-b0ac-09da9c149141/control-plane-machine-set-operator/0.log" Dec 01 09:22:48 crc kubenswrapper[4744]: I1201 09:22:48.695970 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-l9mth_6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc/kube-rbac-proxy/0.log" Dec 01 09:22:48 crc kubenswrapper[4744]: I1201 09:22:48.738986 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-l9mth_6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc/machine-api-operator/0.log" Dec 01 09:23:00 crc kubenswrapper[4744]: I1201 09:23:00.286331 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:23:00 crc kubenswrapper[4744]: E1201 09:23:00.287011 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:23:01 crc kubenswrapper[4744]: I1201 09:23:01.501546 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bjbkb_e05026bf-1db3-4a62-baff-a20dd3d8fd65/cert-manager-controller/0.log" Dec 01 09:23:01 crc kubenswrapper[4744]: I1201 09:23:01.615228 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-bp2pq_5bc4fc9c-b3ed-40f1-b951-8587803ba399/cert-manager-cainjector/0.log" Dec 01 09:23:01 crc kubenswrapper[4744]: I1201 09:23:01.723026 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-w864b_db8c6300-bdd5-4fe7-af23-ed9feb32b844/cert-manager-webhook/0.log" Dec 01 09:23:14 crc kubenswrapper[4744]: I1201 09:23:14.285097 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:23:14 crc kubenswrapper[4744]: E1201 09:23:14.285842 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:23:15 crc kubenswrapper[4744]: I1201 09:23:15.781762 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-vrj4v_9d70fe1c-9023-42af-9b1e-c078b39399f6/nmstate-console-plugin/0.log" Dec 01 09:23:16 crc kubenswrapper[4744]: I1201 09:23:16.004900 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7r6x7_ea5a52f6-b418-4027-b6c0-3d57394d5fa5/nmstate-handler/0.log" Dec 01 09:23:16 crc kubenswrapper[4744]: I1201 09:23:16.088218 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lm574_5cdd7453-8251-4776-82bf-b931ba787d45/kube-rbac-proxy/0.log" Dec 01 09:23:16 crc kubenswrapper[4744]: I1201 09:23:16.142278 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lm574_5cdd7453-8251-4776-82bf-b931ba787d45/nmstate-metrics/0.log" Dec 01 09:23:16 crc kubenswrapper[4744]: I1201 09:23:16.333553 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-mmbdq_ce67e6e1-9203-47bd-abb6-1dfe9b552bca/nmstate-operator/0.log" Dec 01 09:23:16 crc kubenswrapper[4744]: I1201 09:23:16.403628 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-gq68c_5c7e45c3-5458-46ec-8c35-8e9776f6550e/nmstate-webhook/0.log" Dec 01 09:23:26 crc kubenswrapper[4744]: I1201 09:23:26.286471 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:23:26 crc kubenswrapper[4744]: E1201 09:23:26.287148 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:23:30 crc kubenswrapper[4744]: I1201 09:23:30.944834 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bwdwl_edc692c2-6d9e-42fd-b54a-f856dd9f2ed1/kube-rbac-proxy/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.076547 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bwdwl_edc692c2-6d9e-42fd-b54a-f856dd9f2ed1/controller/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.176187 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-c6jzq_481b9c20-e688-4def-a63b-57038302c23c/frr-k8s-webhook-server/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.266330 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.455843 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.495769 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.497372 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.525287 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.661748 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.671821 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.737381 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.758882 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.899539 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.901492 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.924305 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:23:31 crc kubenswrapper[4744]: I1201 09:23:31.992484 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/controller/0.log" Dec 01 09:23:32 crc kubenswrapper[4744]: I1201 09:23:32.094742 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/frr-metrics/0.log" Dec 01 09:23:32 crc kubenswrapper[4744]: I1201 09:23:32.129669 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/kube-rbac-proxy/0.log" Dec 01 09:23:32 crc kubenswrapper[4744]: I1201 09:23:32.237554 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/kube-rbac-proxy-frr/0.log" Dec 01 09:23:32 crc kubenswrapper[4744]: I1201 09:23:32.314373 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/reloader/0.log" Dec 01 09:23:32 crc kubenswrapper[4744]: I1201 09:23:32.491570 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79b5ddfc8f-x7dgg_6d5bd225-d8b4-4836-bb19-7e5c0fd540ae/manager/0.log" Dec 01 09:23:32 crc kubenswrapper[4744]: I1201 09:23:32.607772 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68f89dd57b-kgd5g_84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d/webhook-server/0.log" Dec 01 09:23:32 crc kubenswrapper[4744]: I1201 09:23:32.774561 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6vzqh_3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c/kube-rbac-proxy/0.log" Dec 01 09:23:33 crc kubenswrapper[4744]: I1201 09:23:33.392825 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6vzqh_3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c/speaker/0.log" Dec 01 09:23:33 crc kubenswrapper[4744]: I1201 09:23:33.818542 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/frr/0.log" Dec 01 09:23:38 crc kubenswrapper[4744]: I1201 09:23:38.291684 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:23:38 crc kubenswrapper[4744]: E1201 09:23:38.292526 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.069753 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/util/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.278117 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/util/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.309468 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/pull/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.320317 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/pull/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.499676 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/util/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.620726 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/extract/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.667868 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/util/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.842717 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/pull/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.845435 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/pull/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.863126 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/util/0.log" Dec 01 09:23:45 crc kubenswrapper[4744]: I1201 09:23:45.895692 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/pull/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.043692 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/util/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.054466 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/extract/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.086228 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/pull/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.201268 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/util/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.363117 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/util/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.383950 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/pull/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.426710 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/pull/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.611513 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/util/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.615943 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/pull/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.624732 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/extract/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.775117 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-utilities/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.996881 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-content/0.log" Dec 01 09:23:46 crc kubenswrapper[4744]: I1201 09:23:46.997031 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-utilities/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.004059 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-content/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.160492 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-content/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.165917 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-utilities/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.360806 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-utilities/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.555979 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-content/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.633086 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-utilities/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.652370 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-content/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.886912 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-content/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.898259 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/registry-server/0.log" Dec 01 09:23:47 crc kubenswrapper[4744]: I1201 09:23:47.903489 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-utilities/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.065130 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-99jdl_94e841b2-c9ed-43c3-b96c-07661415a958/marketplace-operator/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.355051 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-utilities/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.610345 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-utilities/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.660123 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-content/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.674558 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/registry-server/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.680639 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-content/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.837939 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-utilities/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.838059 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-content/0.log" Dec 01 09:23:48 crc kubenswrapper[4744]: I1201 09:23:48.890771 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-utilities/0.log" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.020873 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/registry-server/0.log" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.086151 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-utilities/0.log" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.115110 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-content/0.log" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.150987 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-content/0.log" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.284948 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:23:49 crc kubenswrapper[4744]: E1201 09:23:49.285256 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.295090 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-content/0.log" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.311959 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-utilities/0.log" Dec 01 09:23:49 crc kubenswrapper[4744]: I1201 09:23:49.652380 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/registry-server/0.log" Dec 01 09:24:01 crc kubenswrapper[4744]: I1201 09:24:01.642801 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-jt269_d7b03d83-6bcb-4e4d-beb9-d4932e4fab91/prometheus-operator/0.log" Dec 01 09:24:01 crc kubenswrapper[4744]: I1201 09:24:01.824343 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj_50194db7-90e4-4112-83df-f729548b59ca/prometheus-operator-admission-webhook/0.log" Dec 01 09:24:01 crc kubenswrapper[4744]: I1201 09:24:01.914386 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp_92c3c976-90f4-4d13-9d88-348767b50827/prometheus-operator-admission-webhook/0.log" Dec 01 09:24:02 crc kubenswrapper[4744]: I1201 09:24:02.032221 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-m2tkm_1466d56b-9044-4bd8-97f3-02c1feafa96f/operator/0.log" Dec 01 09:24:02 crc kubenswrapper[4744]: I1201 09:24:02.147933 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-tk4rw_d48e874b-8560-49fa-9981-136edeec81f7/perses-operator/0.log" Dec 01 09:24:02 crc kubenswrapper[4744]: I1201 09:24:02.287280 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:24:02 crc kubenswrapper[4744]: E1201 09:24:02.288799 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:24:13 crc kubenswrapper[4744]: I1201 09:24:13.285382 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:24:13 crc kubenswrapper[4744]: E1201 09:24:13.286339 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:24:28 crc kubenswrapper[4744]: I1201 09:24:28.293922 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:24:28 crc kubenswrapper[4744]: I1201 09:24:28.989017 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"5c80c3abc5567fadc0827552b4cdbd5e7a3c17172db5d808070fd49e780b1417"} Dec 01 09:25:29 crc kubenswrapper[4744]: I1201 09:25:29.092848 4744 generic.go:334] "Generic (PLEG): container finished" podID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerID="f7d89553cc1ce987088120c6b33715e47550d0f4f40b4b11460e2f92a07470d4" exitCode=0 Dec 01 09:25:29 crc kubenswrapper[4744]: I1201 09:25:29.092919 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" event={"ID":"26a046ad-bfed-4596-af5c-5c1f8854a826","Type":"ContainerDied","Data":"f7d89553cc1ce987088120c6b33715e47550d0f4f40b4b11460e2f92a07470d4"} Dec 01 09:25:29 crc kubenswrapper[4744]: I1201 09:25:29.094460 4744 scope.go:117] "RemoveContainer" containerID="f7d89553cc1ce987088120c6b33715e47550d0f4f40b4b11460e2f92a07470d4" Dec 01 09:25:29 crc kubenswrapper[4744]: I1201 09:25:29.529185 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4vjqg_must-gather-lwxpd_26a046ad-bfed-4596-af5c-5c1f8854a826/gather/0.log" Dec 01 09:25:37 crc kubenswrapper[4744]: I1201 09:25:37.622767 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4vjqg/must-gather-lwxpd"] Dec 01 09:25:37 crc kubenswrapper[4744]: I1201 09:25:37.623722 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerName="copy" containerID="cri-o://28a0e804cfdeff7ede950f0a07b03984c9dbb6d8b782950fdba615c7845801ed" gracePeriod=2 Dec 01 09:25:37 crc kubenswrapper[4744]: I1201 09:25:37.636398 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4vjqg/must-gather-lwxpd"] Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.177107 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4vjqg_must-gather-lwxpd_26a046ad-bfed-4596-af5c-5c1f8854a826/copy/0.log" Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.177752 4744 generic.go:334] "Generic (PLEG): container finished" podID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerID="28a0e804cfdeff7ede950f0a07b03984c9dbb6d8b782950fdba615c7845801ed" exitCode=143 Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.325973 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4vjqg_must-gather-lwxpd_26a046ad-bfed-4596-af5c-5c1f8854a826/copy/0.log" Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.326623 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.383813 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/26a046ad-bfed-4596-af5c-5c1f8854a826-must-gather-output\") pod \"26a046ad-bfed-4596-af5c-5c1f8854a826\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.383994 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl48t\" (UniqueName: \"kubernetes.io/projected/26a046ad-bfed-4596-af5c-5c1f8854a826-kube-api-access-gl48t\") pod \"26a046ad-bfed-4596-af5c-5c1f8854a826\" (UID: \"26a046ad-bfed-4596-af5c-5c1f8854a826\") " Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.395654 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a046ad-bfed-4596-af5c-5c1f8854a826-kube-api-access-gl48t" (OuterVolumeSpecName: "kube-api-access-gl48t") pod "26a046ad-bfed-4596-af5c-5c1f8854a826" (UID: "26a046ad-bfed-4596-af5c-5c1f8854a826"). InnerVolumeSpecName "kube-api-access-gl48t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.485754 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl48t\" (UniqueName: \"kubernetes.io/projected/26a046ad-bfed-4596-af5c-5c1f8854a826-kube-api-access-gl48t\") on node \"crc\" DevicePath \"\"" Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.534546 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26a046ad-bfed-4596-af5c-5c1f8854a826-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "26a046ad-bfed-4596-af5c-5c1f8854a826" (UID: "26a046ad-bfed-4596-af5c-5c1f8854a826"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:25:38 crc kubenswrapper[4744]: I1201 09:25:38.587206 4744 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/26a046ad-bfed-4596-af5c-5c1f8854a826-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 09:25:39 crc kubenswrapper[4744]: I1201 09:25:39.191580 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4vjqg_must-gather-lwxpd_26a046ad-bfed-4596-af5c-5c1f8854a826/copy/0.log" Dec 01 09:25:39 crc kubenswrapper[4744]: I1201 09:25:39.193134 4744 scope.go:117] "RemoveContainer" containerID="28a0e804cfdeff7ede950f0a07b03984c9dbb6d8b782950fdba615c7845801ed" Dec 01 09:25:39 crc kubenswrapper[4744]: I1201 09:25:39.193308 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4vjqg/must-gather-lwxpd" Dec 01 09:25:39 crc kubenswrapper[4744]: I1201 09:25:39.218461 4744 scope.go:117] "RemoveContainer" containerID="f7d89553cc1ce987088120c6b33715e47550d0f4f40b4b11460e2f92a07470d4" Dec 01 09:25:40 crc kubenswrapper[4744]: I1201 09:25:40.299097 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" path="/var/lib/kubelet/pods/26a046ad-bfed-4596-af5c-5c1f8854a826/volumes" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.085251 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lbtkf"] Dec 01 09:26:28 crc kubenswrapper[4744]: E1201 09:26:28.087010 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="registry-server" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087050 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="registry-server" Dec 01 09:26:28 crc kubenswrapper[4744]: E1201 09:26:28.087080 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerName="copy" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087088 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerName="copy" Dec 01 09:26:28 crc kubenswrapper[4744]: E1201 09:26:28.087102 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerName="gather" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087146 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerName="gather" Dec 01 09:26:28 crc kubenswrapper[4744]: E1201 09:26:28.087171 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="extract-utilities" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087179 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="extract-utilities" Dec 01 09:26:28 crc kubenswrapper[4744]: E1201 09:26:28.087224 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="extract-content" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087236 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="extract-content" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087605 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerName="gather" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087618 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e119ba92-fa29-402c-a60e-fd7027d681bc" containerName="registry-server" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.087636 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a046ad-bfed-4596-af5c-5c1f8854a826" containerName="copy" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.090162 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.095230 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbtkf"] Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.219643 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-catalog-content\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.219698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v55z2\" (UniqueName: \"kubernetes.io/projected/d4e07c30-5b71-4b2a-9c05-240e41653328-kube-api-access-v55z2\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.219806 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-utilities\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.322208 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-catalog-content\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.322255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v55z2\" (UniqueName: \"kubernetes.io/projected/d4e07c30-5b71-4b2a-9c05-240e41653328-kube-api-access-v55z2\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.322336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-utilities\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.322737 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-catalog-content\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.322797 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-utilities\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.341495 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v55z2\" (UniqueName: \"kubernetes.io/projected/d4e07c30-5b71-4b2a-9c05-240e41653328-kube-api-access-v55z2\") pod \"redhat-marketplace-lbtkf\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.448581 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:28 crc kubenswrapper[4744]: I1201 09:26:28.900086 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbtkf"] Dec 01 09:26:29 crc kubenswrapper[4744]: I1201 09:26:29.689630 4744 generic.go:334] "Generic (PLEG): container finished" podID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerID="0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97" exitCode=0 Dec 01 09:26:29 crc kubenswrapper[4744]: I1201 09:26:29.689827 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbtkf" event={"ID":"d4e07c30-5b71-4b2a-9c05-240e41653328","Type":"ContainerDied","Data":"0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97"} Dec 01 09:26:29 crc kubenswrapper[4744]: I1201 09:26:29.689904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbtkf" event={"ID":"d4e07c30-5b71-4b2a-9c05-240e41653328","Type":"ContainerStarted","Data":"5fa4a6ec5f7c57fd8f6794184caa6a119c6025c16c2d341a2b23a6158f9240d9"} Dec 01 09:26:30 crc kubenswrapper[4744]: I1201 09:26:30.700398 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbtkf" event={"ID":"d4e07c30-5b71-4b2a-9c05-240e41653328","Type":"ContainerStarted","Data":"ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd"} Dec 01 09:26:31 crc kubenswrapper[4744]: I1201 09:26:31.710494 4744 generic.go:334] "Generic (PLEG): container finished" podID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerID="ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd" exitCode=0 Dec 01 09:26:31 crc kubenswrapper[4744]: I1201 09:26:31.710561 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbtkf" event={"ID":"d4e07c30-5b71-4b2a-9c05-240e41653328","Type":"ContainerDied","Data":"ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd"} Dec 01 09:26:32 crc kubenswrapper[4744]: I1201 09:26:32.721142 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbtkf" event={"ID":"d4e07c30-5b71-4b2a-9c05-240e41653328","Type":"ContainerStarted","Data":"7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9"} Dec 01 09:26:32 crc kubenswrapper[4744]: I1201 09:26:32.741161 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lbtkf" podStartSLOduration=2.150184707 podStartE2EDuration="4.74113923s" podCreationTimestamp="2025-12-01 09:26:28 +0000 UTC" firstStartedPulling="2025-12-01 09:26:29.691398718 +0000 UTC m=+4121.680456639" lastFinishedPulling="2025-12-01 09:26:32.282353241 +0000 UTC m=+4124.271411162" observedRunningTime="2025-12-01 09:26:32.737071625 +0000 UTC m=+4124.726129546" watchObservedRunningTime="2025-12-01 09:26:32.74113923 +0000 UTC m=+4124.730197161" Dec 01 09:26:38 crc kubenswrapper[4744]: I1201 09:26:38.448886 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:38 crc kubenswrapper[4744]: I1201 09:26:38.449288 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:38 crc kubenswrapper[4744]: I1201 09:26:38.494684 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:38 crc kubenswrapper[4744]: I1201 09:26:38.830176 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:38 crc kubenswrapper[4744]: I1201 09:26:38.878195 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbtkf"] Dec 01 09:26:40 crc kubenswrapper[4744]: I1201 09:26:40.791625 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lbtkf" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="registry-server" containerID="cri-o://7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9" gracePeriod=2 Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.257728 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.281049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-catalog-content\") pod \"d4e07c30-5b71-4b2a-9c05-240e41653328\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.281106 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-utilities\") pod \"d4e07c30-5b71-4b2a-9c05-240e41653328\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.281351 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v55z2\" (UniqueName: \"kubernetes.io/projected/d4e07c30-5b71-4b2a-9c05-240e41653328-kube-api-access-v55z2\") pod \"d4e07c30-5b71-4b2a-9c05-240e41653328\" (UID: \"d4e07c30-5b71-4b2a-9c05-240e41653328\") " Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.282119 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-utilities" (OuterVolumeSpecName: "utilities") pod "d4e07c30-5b71-4b2a-9c05-240e41653328" (UID: "d4e07c30-5b71-4b2a-9c05-240e41653328"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.292103 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e07c30-5b71-4b2a-9c05-240e41653328-kube-api-access-v55z2" (OuterVolumeSpecName: "kube-api-access-v55z2") pod "d4e07c30-5b71-4b2a-9c05-240e41653328" (UID: "d4e07c30-5b71-4b2a-9c05-240e41653328"). InnerVolumeSpecName "kube-api-access-v55z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.303430 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4e07c30-5b71-4b2a-9c05-240e41653328" (UID: "d4e07c30-5b71-4b2a-9c05-240e41653328"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.383777 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v55z2\" (UniqueName: \"kubernetes.io/projected/d4e07c30-5b71-4b2a-9c05-240e41653328-kube-api-access-v55z2\") on node \"crc\" DevicePath \"\"" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.384038 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.384055 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4e07c30-5b71-4b2a-9c05-240e41653328-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.801812 4744 generic.go:334] "Generic (PLEG): container finished" podID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerID="7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9" exitCode=0 Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.801855 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbtkf" event={"ID":"d4e07c30-5b71-4b2a-9c05-240e41653328","Type":"ContainerDied","Data":"7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9"} Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.801882 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbtkf" event={"ID":"d4e07c30-5b71-4b2a-9c05-240e41653328","Type":"ContainerDied","Data":"5fa4a6ec5f7c57fd8f6794184caa6a119c6025c16c2d341a2b23a6158f9240d9"} Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.801899 4744 scope.go:117] "RemoveContainer" containerID="7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.802008 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbtkf" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.825846 4744 scope.go:117] "RemoveContainer" containerID="ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.851187 4744 scope.go:117] "RemoveContainer" containerID="0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.859380 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbtkf"] Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.872767 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbtkf"] Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.907042 4744 scope.go:117] "RemoveContainer" containerID="7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9" Dec 01 09:26:41 crc kubenswrapper[4744]: E1201 09:26:41.907607 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9\": container with ID starting with 7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9 not found: ID does not exist" containerID="7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.907668 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9"} err="failed to get container status \"7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9\": rpc error: code = NotFound desc = could not find container \"7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9\": container with ID starting with 7d56f874b8815f900fdc55902a0972f1197b670549a1e9b02e0a28ceb46882b9 not found: ID does not exist" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.907697 4744 scope.go:117] "RemoveContainer" containerID="ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd" Dec 01 09:26:41 crc kubenswrapper[4744]: E1201 09:26:41.908075 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd\": container with ID starting with ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd not found: ID does not exist" containerID="ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.908109 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd"} err="failed to get container status \"ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd\": rpc error: code = NotFound desc = could not find container \"ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd\": container with ID starting with ee5098be0ad4f3a7529c054d4062ad34fc6d10558e585abe3734a87db06be5cd not found: ID does not exist" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.908136 4744 scope.go:117] "RemoveContainer" containerID="0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97" Dec 01 09:26:41 crc kubenswrapper[4744]: E1201 09:26:41.908481 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97\": container with ID starting with 0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97 not found: ID does not exist" containerID="0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97" Dec 01 09:26:41 crc kubenswrapper[4744]: I1201 09:26:41.908502 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97"} err="failed to get container status \"0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97\": rpc error: code = NotFound desc = could not find container \"0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97\": container with ID starting with 0f77b10a844a5297ec677824011841c123741bdbaa4ecc974849db9fffed1c97 not found: ID does not exist" Dec 01 09:26:42 crc kubenswrapper[4744]: I1201 09:26:42.298230 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" path="/var/lib/kubelet/pods/d4e07c30-5b71-4b2a-9c05-240e41653328/volumes" Dec 01 09:26:48 crc kubenswrapper[4744]: I1201 09:26:48.635680 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:26:48 crc kubenswrapper[4744]: I1201 09:26:48.636676 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:27:04 crc kubenswrapper[4744]: I1201 09:27:04.896628 4744 scope.go:117] "RemoveContainer" containerID="755e0d4e87fce0416921a826de0cdfe4c2083dd65a87bcf9e522997055aff25d" Dec 01 09:27:18 crc kubenswrapper[4744]: I1201 09:27:18.634272 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:27:18 crc kubenswrapper[4744]: I1201 09:27:18.635842 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.318968 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4rvsg"] Dec 01 09:27:33 crc kubenswrapper[4744]: E1201 09:27:33.320129 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="extract-utilities" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.320223 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="extract-utilities" Dec 01 09:27:33 crc kubenswrapper[4744]: E1201 09:27:33.320263 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="registry-server" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.320270 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="registry-server" Dec 01 09:27:33 crc kubenswrapper[4744]: E1201 09:27:33.320284 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="extract-content" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.320292 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="extract-content" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.320606 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4e07c30-5b71-4b2a-9c05-240e41653328" containerName="registry-server" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.322594 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.332628 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4rvsg"] Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.425247 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k894b\" (UniqueName: \"kubernetes.io/projected/38ca24ad-4709-4353-af8d-95533ea584c1-kube-api-access-k894b\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.425313 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-utilities\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.425571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-catalog-content\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.529565 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k894b\" (UniqueName: \"kubernetes.io/projected/38ca24ad-4709-4353-af8d-95533ea584c1-kube-api-access-k894b\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.529740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-utilities\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.529958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-catalog-content\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.530334 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-utilities\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.530824 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-catalog-content\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.551090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k894b\" (UniqueName: \"kubernetes.io/projected/38ca24ad-4709-4353-af8d-95533ea584c1-kube-api-access-k894b\") pod \"redhat-operators-4rvsg\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:33 crc kubenswrapper[4744]: I1201 09:27:33.658242 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:34 crc kubenswrapper[4744]: I1201 09:27:34.231489 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4rvsg"] Dec 01 09:27:34 crc kubenswrapper[4744]: I1201 09:27:34.323188 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rvsg" event={"ID":"38ca24ad-4709-4353-af8d-95533ea584c1","Type":"ContainerStarted","Data":"28bf99f0a7b5fbd14d246d6206d05c568ebd1c4b146ce3921933bd528f1f1af3"} Dec 01 09:27:35 crc kubenswrapper[4744]: I1201 09:27:35.332940 4744 generic.go:334] "Generic (PLEG): container finished" podID="38ca24ad-4709-4353-af8d-95533ea584c1" containerID="df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1" exitCode=0 Dec 01 09:27:35 crc kubenswrapper[4744]: I1201 09:27:35.333228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rvsg" event={"ID":"38ca24ad-4709-4353-af8d-95533ea584c1","Type":"ContainerDied","Data":"df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1"} Dec 01 09:27:35 crc kubenswrapper[4744]: I1201 09:27:35.336404 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.716626 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c6hv6"] Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.719597 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.727138 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c6hv6"] Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.798581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqtls\" (UniqueName: \"kubernetes.io/projected/a7eb0839-93ee-4188-83ac-e23ab25e537a-kube-api-access-qqtls\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.798730 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-catalog-content\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.798897 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-utilities\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.900886 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqtls\" (UniqueName: \"kubernetes.io/projected/a7eb0839-93ee-4188-83ac-e23ab25e537a-kube-api-access-qqtls\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.901060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-catalog-content\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.901129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-utilities\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.901617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-catalog-content\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.901655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-utilities\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:36 crc kubenswrapper[4744]: I1201 09:27:36.931348 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqtls\" (UniqueName: \"kubernetes.io/projected/a7eb0839-93ee-4188-83ac-e23ab25e537a-kube-api-access-qqtls\") pod \"certified-operators-c6hv6\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:37 crc kubenswrapper[4744]: I1201 09:27:37.038158 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:37 crc kubenswrapper[4744]: I1201 09:27:37.358309 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rvsg" event={"ID":"38ca24ad-4709-4353-af8d-95533ea584c1","Type":"ContainerStarted","Data":"73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1"} Dec 01 09:27:37 crc kubenswrapper[4744]: I1201 09:27:37.558182 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c6hv6"] Dec 01 09:27:38 crc kubenswrapper[4744]: I1201 09:27:38.382137 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerID="2e9decf918bf945e13f4a3b6c38b927dfd65e5f3c9332f5ce93bf0379812c9ba" exitCode=0 Dec 01 09:27:38 crc kubenswrapper[4744]: I1201 09:27:38.382586 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6hv6" event={"ID":"a7eb0839-93ee-4188-83ac-e23ab25e537a","Type":"ContainerDied","Data":"2e9decf918bf945e13f4a3b6c38b927dfd65e5f3c9332f5ce93bf0379812c9ba"} Dec 01 09:27:38 crc kubenswrapper[4744]: I1201 09:27:38.382663 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6hv6" event={"ID":"a7eb0839-93ee-4188-83ac-e23ab25e537a","Type":"ContainerStarted","Data":"f11eed1b315ebd08b1113716876f693104a4b1605c101c17661ee84f2813b7e2"} Dec 01 09:27:39 crc kubenswrapper[4744]: I1201 09:27:39.398116 4744 generic.go:334] "Generic (PLEG): container finished" podID="38ca24ad-4709-4353-af8d-95533ea584c1" containerID="73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1" exitCode=0 Dec 01 09:27:39 crc kubenswrapper[4744]: I1201 09:27:39.398486 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rvsg" event={"ID":"38ca24ad-4709-4353-af8d-95533ea584c1","Type":"ContainerDied","Data":"73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1"} Dec 01 09:27:40 crc kubenswrapper[4744]: I1201 09:27:40.408392 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerID="fdacc725fbe1e18841e6dc676e8c6b52e887afe2b0c7f1bd148577d53435b62e" exitCode=0 Dec 01 09:27:40 crc kubenswrapper[4744]: I1201 09:27:40.408446 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6hv6" event={"ID":"a7eb0839-93ee-4188-83ac-e23ab25e537a","Type":"ContainerDied","Data":"fdacc725fbe1e18841e6dc676e8c6b52e887afe2b0c7f1bd148577d53435b62e"} Dec 01 09:27:40 crc kubenswrapper[4744]: I1201 09:27:40.411247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rvsg" event={"ID":"38ca24ad-4709-4353-af8d-95533ea584c1","Type":"ContainerStarted","Data":"90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5"} Dec 01 09:27:40 crc kubenswrapper[4744]: I1201 09:27:40.453023 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4rvsg" podStartSLOduration=2.877256609 podStartE2EDuration="7.453004202s" podCreationTimestamp="2025-12-01 09:27:33 +0000 UTC" firstStartedPulling="2025-12-01 09:27:35.33609411 +0000 UTC m=+4187.325152041" lastFinishedPulling="2025-12-01 09:27:39.911841713 +0000 UTC m=+4191.900899634" observedRunningTime="2025-12-01 09:27:40.44449272 +0000 UTC m=+4192.433550671" watchObservedRunningTime="2025-12-01 09:27:40.453004202 +0000 UTC m=+4192.442062123" Dec 01 09:27:42 crc kubenswrapper[4744]: I1201 09:27:42.432363 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6hv6" event={"ID":"a7eb0839-93ee-4188-83ac-e23ab25e537a","Type":"ContainerStarted","Data":"c860f7ca32eb09ce5eb8de5136c1d5a3078db250676188550ded140901c5f320"} Dec 01 09:27:42 crc kubenswrapper[4744]: I1201 09:27:42.454594 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c6hv6" podStartSLOduration=3.593357467 podStartE2EDuration="6.454573776s" podCreationTimestamp="2025-12-01 09:27:36 +0000 UTC" firstStartedPulling="2025-12-01 09:27:38.385097623 +0000 UTC m=+4190.374155554" lastFinishedPulling="2025-12-01 09:27:41.246313942 +0000 UTC m=+4193.235371863" observedRunningTime="2025-12-01 09:27:42.449756309 +0000 UTC m=+4194.438814240" watchObservedRunningTime="2025-12-01 09:27:42.454573776 +0000 UTC m=+4194.443631707" Dec 01 09:27:43 crc kubenswrapper[4744]: I1201 09:27:43.658735 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:43 crc kubenswrapper[4744]: I1201 09:27:43.658804 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:44 crc kubenswrapper[4744]: I1201 09:27:44.710897 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4rvsg" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="registry-server" probeResult="failure" output=< Dec 01 09:27:44 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Dec 01 09:27:44 crc kubenswrapper[4744]: > Dec 01 09:27:47 crc kubenswrapper[4744]: I1201 09:27:47.038347 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:47 crc kubenswrapper[4744]: I1201 09:27:47.039637 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:47 crc kubenswrapper[4744]: I1201 09:27:47.088510 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:47 crc kubenswrapper[4744]: I1201 09:27:47.520137 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:48 crc kubenswrapper[4744]: I1201 09:27:48.633880 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:27:48 crc kubenswrapper[4744]: I1201 09:27:48.634240 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:27:48 crc kubenswrapper[4744]: I1201 09:27:48.634351 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 09:27:48 crc kubenswrapper[4744]: I1201 09:27:48.635513 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c80c3abc5567fadc0827552b4cdbd5e7a3c17172db5d808070fd49e780b1417"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:27:48 crc kubenswrapper[4744]: I1201 09:27:48.635619 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://5c80c3abc5567fadc0827552b4cdbd5e7a3c17172db5d808070fd49e780b1417" gracePeriod=600 Dec 01 09:27:49 crc kubenswrapper[4744]: I1201 09:27:49.495028 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="5c80c3abc5567fadc0827552b4cdbd5e7a3c17172db5d808070fd49e780b1417" exitCode=0 Dec 01 09:27:49 crc kubenswrapper[4744]: I1201 09:27:49.495084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"5c80c3abc5567fadc0827552b4cdbd5e7a3c17172db5d808070fd49e780b1417"} Dec 01 09:27:49 crc kubenswrapper[4744]: I1201 09:27:49.495356 4744 scope.go:117] "RemoveContainer" containerID="b3630e20ac9831157b997e914458ed61d1f22c45843bd4fa4abe3fd48c83679e" Dec 01 09:27:50 crc kubenswrapper[4744]: I1201 09:27:50.507688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353"} Dec 01 09:27:50 crc kubenswrapper[4744]: I1201 09:27:50.703890 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c6hv6"] Dec 01 09:27:50 crc kubenswrapper[4744]: I1201 09:27:50.704099 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c6hv6" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="registry-server" containerID="cri-o://c860f7ca32eb09ce5eb8de5136c1d5a3078db250676188550ded140901c5f320" gracePeriod=2 Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.521941 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerID="c860f7ca32eb09ce5eb8de5136c1d5a3078db250676188550ded140901c5f320" exitCode=0 Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.522012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6hv6" event={"ID":"a7eb0839-93ee-4188-83ac-e23ab25e537a","Type":"ContainerDied","Data":"c860f7ca32eb09ce5eb8de5136c1d5a3078db250676188550ded140901c5f320"} Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.522342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6hv6" event={"ID":"a7eb0839-93ee-4188-83ac-e23ab25e537a","Type":"ContainerDied","Data":"f11eed1b315ebd08b1113716876f693104a4b1605c101c17661ee84f2813b7e2"} Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.522368 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f11eed1b315ebd08b1113716876f693104a4b1605c101c17661ee84f2813b7e2" Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.763202 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.828079 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-catalog-content\") pod \"a7eb0839-93ee-4188-83ac-e23ab25e537a\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.828264 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-utilities\") pod \"a7eb0839-93ee-4188-83ac-e23ab25e537a\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.828308 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqtls\" (UniqueName: \"kubernetes.io/projected/a7eb0839-93ee-4188-83ac-e23ab25e537a-kube-api-access-qqtls\") pod \"a7eb0839-93ee-4188-83ac-e23ab25e537a\" (UID: \"a7eb0839-93ee-4188-83ac-e23ab25e537a\") " Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.829995 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-utilities" (OuterVolumeSpecName: "utilities") pod "a7eb0839-93ee-4188-83ac-e23ab25e537a" (UID: "a7eb0839-93ee-4188-83ac-e23ab25e537a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.834547 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7eb0839-93ee-4188-83ac-e23ab25e537a-kube-api-access-qqtls" (OuterVolumeSpecName: "kube-api-access-qqtls") pod "a7eb0839-93ee-4188-83ac-e23ab25e537a" (UID: "a7eb0839-93ee-4188-83ac-e23ab25e537a"). InnerVolumeSpecName "kube-api-access-qqtls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.900855 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7eb0839-93ee-4188-83ac-e23ab25e537a" (UID: "a7eb0839-93ee-4188-83ac-e23ab25e537a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.931107 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.931170 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7eb0839-93ee-4188-83ac-e23ab25e537a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:51 crc kubenswrapper[4744]: I1201 09:27:51.931182 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqtls\" (UniqueName: \"kubernetes.io/projected/a7eb0839-93ee-4188-83ac-e23ab25e537a-kube-api-access-qqtls\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:52 crc kubenswrapper[4744]: I1201 09:27:52.531383 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6hv6" Dec 01 09:27:52 crc kubenswrapper[4744]: I1201 09:27:52.557634 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c6hv6"] Dec 01 09:27:52 crc kubenswrapper[4744]: I1201 09:27:52.566972 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c6hv6"] Dec 01 09:27:53 crc kubenswrapper[4744]: I1201 09:27:53.710816 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:53 crc kubenswrapper[4744]: I1201 09:27:53.764678 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:54 crc kubenswrapper[4744]: I1201 09:27:54.300600 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" path="/var/lib/kubelet/pods/a7eb0839-93ee-4188-83ac-e23ab25e537a/volumes" Dec 01 09:27:54 crc kubenswrapper[4744]: I1201 09:27:54.902329 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4rvsg"] Dec 01 09:27:55 crc kubenswrapper[4744]: I1201 09:27:55.563252 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4rvsg" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="registry-server" containerID="cri-o://90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5" gracePeriod=2 Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.041013 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.107514 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k894b\" (UniqueName: \"kubernetes.io/projected/38ca24ad-4709-4353-af8d-95533ea584c1-kube-api-access-k894b\") pod \"38ca24ad-4709-4353-af8d-95533ea584c1\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.107668 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-catalog-content\") pod \"38ca24ad-4709-4353-af8d-95533ea584c1\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.107816 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-utilities\") pod \"38ca24ad-4709-4353-af8d-95533ea584c1\" (UID: \"38ca24ad-4709-4353-af8d-95533ea584c1\") " Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.108657 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-utilities" (OuterVolumeSpecName: "utilities") pod "38ca24ad-4709-4353-af8d-95533ea584c1" (UID: "38ca24ad-4709-4353-af8d-95533ea584c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.113593 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ca24ad-4709-4353-af8d-95533ea584c1-kube-api-access-k894b" (OuterVolumeSpecName: "kube-api-access-k894b") pod "38ca24ad-4709-4353-af8d-95533ea584c1" (UID: "38ca24ad-4709-4353-af8d-95533ea584c1"). InnerVolumeSpecName "kube-api-access-k894b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.209787 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.209824 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k894b\" (UniqueName: \"kubernetes.io/projected/38ca24ad-4709-4353-af8d-95533ea584c1-kube-api-access-k894b\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.217774 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38ca24ad-4709-4353-af8d-95533ea584c1" (UID: "38ca24ad-4709-4353-af8d-95533ea584c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.311640 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38ca24ad-4709-4353-af8d-95533ea584c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.572816 4744 generic.go:334] "Generic (PLEG): container finished" podID="38ca24ad-4709-4353-af8d-95533ea584c1" containerID="90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5" exitCode=0 Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.572857 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rvsg" event={"ID":"38ca24ad-4709-4353-af8d-95533ea584c1","Type":"ContainerDied","Data":"90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5"} Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.573177 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rvsg" event={"ID":"38ca24ad-4709-4353-af8d-95533ea584c1","Type":"ContainerDied","Data":"28bf99f0a7b5fbd14d246d6206d05c568ebd1c4b146ce3921933bd528f1f1af3"} Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.573199 4744 scope.go:117] "RemoveContainer" containerID="90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.572874 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rvsg" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.595761 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4rvsg"] Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.604353 4744 scope.go:117] "RemoveContainer" containerID="73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.605418 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4rvsg"] Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.643335 4744 scope.go:117] "RemoveContainer" containerID="df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.684189 4744 scope.go:117] "RemoveContainer" containerID="90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5" Dec 01 09:27:56 crc kubenswrapper[4744]: E1201 09:27:56.689401 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5\": container with ID starting with 90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5 not found: ID does not exist" containerID="90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.689475 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5"} err="failed to get container status \"90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5\": rpc error: code = NotFound desc = could not find container \"90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5\": container with ID starting with 90c96049e5820cdca07bec7e542364e735ec2479e5090ee07ee4935372dff3c5 not found: ID does not exist" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.689507 4744 scope.go:117] "RemoveContainer" containerID="73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1" Dec 01 09:27:56 crc kubenswrapper[4744]: E1201 09:27:56.691019 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1\": container with ID starting with 73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1 not found: ID does not exist" containerID="73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.691075 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1"} err="failed to get container status \"73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1\": rpc error: code = NotFound desc = could not find container \"73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1\": container with ID starting with 73a3ff13226c69b98fe63d6a747783d21eb779561ab91e57bebe344510f636f1 not found: ID does not exist" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.691106 4744 scope.go:117] "RemoveContainer" containerID="df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1" Dec 01 09:27:56 crc kubenswrapper[4744]: E1201 09:27:56.693140 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1\": container with ID starting with df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1 not found: ID does not exist" containerID="df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1" Dec 01 09:27:56 crc kubenswrapper[4744]: I1201 09:27:56.693174 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1"} err="failed to get container status \"df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1\": rpc error: code = NotFound desc = could not find container \"df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1\": container with ID starting with df77c9df70bf6b990fe2c49d5c8d17b430f792f8c9a2e7e1f326d682198a4da1 not found: ID does not exist" Dec 01 09:27:58 crc kubenswrapper[4744]: I1201 09:27:58.297706 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" path="/var/lib/kubelet/pods/38ca24ad-4709-4353-af8d-95533ea584c1/volumes" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.756186 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x4cr8/must-gather-kg98g"] Dec 01 09:28:12 crc kubenswrapper[4744]: E1201 09:28:12.757324 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="extract-utilities" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757338 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="extract-utilities" Dec 01 09:28:12 crc kubenswrapper[4744]: E1201 09:28:12.757355 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="registry-server" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757361 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="registry-server" Dec 01 09:28:12 crc kubenswrapper[4744]: E1201 09:28:12.757375 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="extract-utilities" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757382 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="extract-utilities" Dec 01 09:28:12 crc kubenswrapper[4744]: E1201 09:28:12.757420 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="extract-content" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757428 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="extract-content" Dec 01 09:28:12 crc kubenswrapper[4744]: E1201 09:28:12.757437 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="registry-server" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757446 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="registry-server" Dec 01 09:28:12 crc kubenswrapper[4744]: E1201 09:28:12.757465 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="extract-content" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757472 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="extract-content" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757688 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7eb0839-93ee-4188-83ac-e23ab25e537a" containerName="registry-server" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.757721 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ca24ad-4709-4353-af8d-95533ea584c1" containerName="registry-server" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.758965 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.764129 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x4cr8"/"kube-root-ca.crt" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.766339 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x4cr8"/"openshift-service-ca.crt" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.768711 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x4cr8/must-gather-kg98g"] Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.835442 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9rbk\" (UniqueName: \"kubernetes.io/projected/4caf2e95-e84d-4818-bdea-651a449165fa-kube-api-access-w9rbk\") pod \"must-gather-kg98g\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.835646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4caf2e95-e84d-4818-bdea-651a449165fa-must-gather-output\") pod \"must-gather-kg98g\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.936945 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9rbk\" (UniqueName: \"kubernetes.io/projected/4caf2e95-e84d-4818-bdea-651a449165fa-kube-api-access-w9rbk\") pod \"must-gather-kg98g\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.937112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4caf2e95-e84d-4818-bdea-651a449165fa-must-gather-output\") pod \"must-gather-kg98g\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.937552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4caf2e95-e84d-4818-bdea-651a449165fa-must-gather-output\") pod \"must-gather-kg98g\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:12 crc kubenswrapper[4744]: I1201 09:28:12.960742 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9rbk\" (UniqueName: \"kubernetes.io/projected/4caf2e95-e84d-4818-bdea-651a449165fa-kube-api-access-w9rbk\") pod \"must-gather-kg98g\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:13 crc kubenswrapper[4744]: I1201 09:28:13.095893 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:28:13 crc kubenswrapper[4744]: I1201 09:28:13.588733 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x4cr8/must-gather-kg98g"] Dec 01 09:28:13 crc kubenswrapper[4744]: I1201 09:28:13.759206 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/must-gather-kg98g" event={"ID":"4caf2e95-e84d-4818-bdea-651a449165fa","Type":"ContainerStarted","Data":"ac77b0cbf2a838582e8b29e22498d721117fa877b3e233da1f7108d9cf9f3d1a"} Dec 01 09:28:14 crc kubenswrapper[4744]: I1201 09:28:14.771469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/must-gather-kg98g" event={"ID":"4caf2e95-e84d-4818-bdea-651a449165fa","Type":"ContainerStarted","Data":"b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa"} Dec 01 09:28:14 crc kubenswrapper[4744]: I1201 09:28:14.771817 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/must-gather-kg98g" event={"ID":"4caf2e95-e84d-4818-bdea-651a449165fa","Type":"ContainerStarted","Data":"1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0"} Dec 01 09:28:14 crc kubenswrapper[4744]: I1201 09:28:14.796479 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x4cr8/must-gather-kg98g" podStartSLOduration=2.796462297 podStartE2EDuration="2.796462297s" podCreationTimestamp="2025-12-01 09:28:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:28:14.793946156 +0000 UTC m=+4226.783004097" watchObservedRunningTime="2025-12-01 09:28:14.796462297 +0000 UTC m=+4226.785520218" Dec 01 09:28:17 crc kubenswrapper[4744]: I1201 09:28:17.793462 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x4cr8/crc-debug-5p55d"] Dec 01 09:28:17 crc kubenswrapper[4744]: I1201 09:28:17.795302 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:17 crc kubenswrapper[4744]: I1201 09:28:17.799227 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x4cr8"/"default-dockercfg-fb7cd" Dec 01 09:28:17 crc kubenswrapper[4744]: I1201 09:28:17.947076 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpxvt\" (UniqueName: \"kubernetes.io/projected/cda5b688-fdea-4b3f-8696-1b1c5edd6069-kube-api-access-jpxvt\") pod \"crc-debug-5p55d\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:17 crc kubenswrapper[4744]: I1201 09:28:17.947159 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cda5b688-fdea-4b3f-8696-1b1c5edd6069-host\") pod \"crc-debug-5p55d\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.049756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpxvt\" (UniqueName: \"kubernetes.io/projected/cda5b688-fdea-4b3f-8696-1b1c5edd6069-kube-api-access-jpxvt\") pod \"crc-debug-5p55d\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.049828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cda5b688-fdea-4b3f-8696-1b1c5edd6069-host\") pod \"crc-debug-5p55d\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.050084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cda5b688-fdea-4b3f-8696-1b1c5edd6069-host\") pod \"crc-debug-5p55d\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.075506 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpxvt\" (UniqueName: \"kubernetes.io/projected/cda5b688-fdea-4b3f-8696-1b1c5edd6069-kube-api-access-jpxvt\") pod \"crc-debug-5p55d\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.116648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:18 crc kubenswrapper[4744]: W1201 09:28:18.146226 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcda5b688_fdea_4b3f_8696_1b1c5edd6069.slice/crio-9262fd217b471a61177c52bdbd7eb5d30eabf2c6ac9025882a965cac744d5db2 WatchSource:0}: Error finding container 9262fd217b471a61177c52bdbd7eb5d30eabf2c6ac9025882a965cac744d5db2: Status 404 returned error can't find the container with id 9262fd217b471a61177c52bdbd7eb5d30eabf2c6ac9025882a965cac744d5db2 Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.814844 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" event={"ID":"cda5b688-fdea-4b3f-8696-1b1c5edd6069","Type":"ContainerStarted","Data":"719f75a638cb37590832179ee14f9e648d86141df911dc4246343224dd15f15f"} Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.815390 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" event={"ID":"cda5b688-fdea-4b3f-8696-1b1c5edd6069","Type":"ContainerStarted","Data":"9262fd217b471a61177c52bdbd7eb5d30eabf2c6ac9025882a965cac744d5db2"} Dec 01 09:28:18 crc kubenswrapper[4744]: I1201 09:28:18.835874 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" podStartSLOduration=1.8358567460000002 podStartE2EDuration="1.835856746s" podCreationTimestamp="2025-12-01 09:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:28:18.830505754 +0000 UTC m=+4230.819563685" watchObservedRunningTime="2025-12-01 09:28:18.835856746 +0000 UTC m=+4230.824914667" Dec 01 09:28:30 crc kubenswrapper[4744]: I1201 09:28:30.940866 4744 generic.go:334] "Generic (PLEG): container finished" podID="cda5b688-fdea-4b3f-8696-1b1c5edd6069" containerID="719f75a638cb37590832179ee14f9e648d86141df911dc4246343224dd15f15f" exitCode=0 Dec 01 09:28:30 crc kubenswrapper[4744]: I1201 09:28:30.940978 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" event={"ID":"cda5b688-fdea-4b3f-8696-1b1c5edd6069","Type":"ContainerDied","Data":"719f75a638cb37590832179ee14f9e648d86141df911dc4246343224dd15f15f"} Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.063451 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.099784 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x4cr8/crc-debug-5p55d"] Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.111078 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x4cr8/crc-debug-5p55d"] Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.205330 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpxvt\" (UniqueName: \"kubernetes.io/projected/cda5b688-fdea-4b3f-8696-1b1c5edd6069-kube-api-access-jpxvt\") pod \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.205568 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cda5b688-fdea-4b3f-8696-1b1c5edd6069-host\") pod \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\" (UID: \"cda5b688-fdea-4b3f-8696-1b1c5edd6069\") " Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.205787 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cda5b688-fdea-4b3f-8696-1b1c5edd6069-host" (OuterVolumeSpecName: "host") pod "cda5b688-fdea-4b3f-8696-1b1c5edd6069" (UID: "cda5b688-fdea-4b3f-8696-1b1c5edd6069"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.206145 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cda5b688-fdea-4b3f-8696-1b1c5edd6069-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.214682 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cda5b688-fdea-4b3f-8696-1b1c5edd6069-kube-api-access-jpxvt" (OuterVolumeSpecName: "kube-api-access-jpxvt") pod "cda5b688-fdea-4b3f-8696-1b1c5edd6069" (UID: "cda5b688-fdea-4b3f-8696-1b1c5edd6069"). InnerVolumeSpecName "kube-api-access-jpxvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.299479 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cda5b688-fdea-4b3f-8696-1b1c5edd6069" path="/var/lib/kubelet/pods/cda5b688-fdea-4b3f-8696-1b1c5edd6069/volumes" Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.307882 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpxvt\" (UniqueName: \"kubernetes.io/projected/cda5b688-fdea-4b3f-8696-1b1c5edd6069-kube-api-access-jpxvt\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.958561 4744 scope.go:117] "RemoveContainer" containerID="719f75a638cb37590832179ee14f9e648d86141df911dc4246343224dd15f15f" Dec 01 09:28:32 crc kubenswrapper[4744]: I1201 09:28:32.958594 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-5p55d" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.286781 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x4cr8/crc-debug-vzq7b"] Dec 01 09:28:33 crc kubenswrapper[4744]: E1201 09:28:33.287170 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cda5b688-fdea-4b3f-8696-1b1c5edd6069" containerName="container-00" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.287184 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cda5b688-fdea-4b3f-8696-1b1c5edd6069" containerName="container-00" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.287427 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cda5b688-fdea-4b3f-8696-1b1c5edd6069" containerName="container-00" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.288240 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.290994 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x4cr8"/"default-dockercfg-fb7cd" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.431153 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4f96cd1-f706-4c28-8437-4d54eb138bd7-host\") pod \"crc-debug-vzq7b\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.431215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6smj\" (UniqueName: \"kubernetes.io/projected/c4f96cd1-f706-4c28-8437-4d54eb138bd7-kube-api-access-p6smj\") pod \"crc-debug-vzq7b\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.533025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4f96cd1-f706-4c28-8437-4d54eb138bd7-host\") pod \"crc-debug-vzq7b\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.533076 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6smj\" (UniqueName: \"kubernetes.io/projected/c4f96cd1-f706-4c28-8437-4d54eb138bd7-kube-api-access-p6smj\") pod \"crc-debug-vzq7b\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.533157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4f96cd1-f706-4c28-8437-4d54eb138bd7-host\") pod \"crc-debug-vzq7b\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.553852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6smj\" (UniqueName: \"kubernetes.io/projected/c4f96cd1-f706-4c28-8437-4d54eb138bd7-kube-api-access-p6smj\") pod \"crc-debug-vzq7b\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.606935 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:33 crc kubenswrapper[4744]: I1201 09:28:33.968071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" event={"ID":"c4f96cd1-f706-4c28-8437-4d54eb138bd7","Type":"ContainerStarted","Data":"3a9053192525f427f19e1802ca33e46991fdf36fba21dd7dea0bd5577298d736"} Dec 01 09:28:34 crc kubenswrapper[4744]: I1201 09:28:34.980458 4744 generic.go:334] "Generic (PLEG): container finished" podID="c4f96cd1-f706-4c28-8437-4d54eb138bd7" containerID="e5935e70c30cab11e38e2f1f2ba798c26ac4416de72350dad59db62ac3aa51ae" exitCode=1 Dec 01 09:28:34 crc kubenswrapper[4744]: I1201 09:28:34.980528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" event={"ID":"c4f96cd1-f706-4c28-8437-4d54eb138bd7","Type":"ContainerDied","Data":"e5935e70c30cab11e38e2f1f2ba798c26ac4416de72350dad59db62ac3aa51ae"} Dec 01 09:28:35 crc kubenswrapper[4744]: I1201 09:28:35.026333 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x4cr8/crc-debug-vzq7b"] Dec 01 09:28:35 crc kubenswrapper[4744]: I1201 09:28:35.042467 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x4cr8/crc-debug-vzq7b"] Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.109466 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.180327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4f96cd1-f706-4c28-8437-4d54eb138bd7-host\") pod \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.180488 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c4f96cd1-f706-4c28-8437-4d54eb138bd7-host" (OuterVolumeSpecName: "host") pod "c4f96cd1-f706-4c28-8437-4d54eb138bd7" (UID: "c4f96cd1-f706-4c28-8437-4d54eb138bd7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.180570 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6smj\" (UniqueName: \"kubernetes.io/projected/c4f96cd1-f706-4c28-8437-4d54eb138bd7-kube-api-access-p6smj\") pod \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\" (UID: \"c4f96cd1-f706-4c28-8437-4d54eb138bd7\") " Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.181133 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4f96cd1-f706-4c28-8437-4d54eb138bd7-host\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.196230 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f96cd1-f706-4c28-8437-4d54eb138bd7-kube-api-access-p6smj" (OuterVolumeSpecName: "kube-api-access-p6smj") pod "c4f96cd1-f706-4c28-8437-4d54eb138bd7" (UID: "c4f96cd1-f706-4c28-8437-4d54eb138bd7"). InnerVolumeSpecName "kube-api-access-p6smj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.283040 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6smj\" (UniqueName: \"kubernetes.io/projected/c4f96cd1-f706-4c28-8437-4d54eb138bd7-kube-api-access-p6smj\") on node \"crc\" DevicePath \"\"" Dec 01 09:28:36 crc kubenswrapper[4744]: I1201 09:28:36.295232 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f96cd1-f706-4c28-8437-4d54eb138bd7" path="/var/lib/kubelet/pods/c4f96cd1-f706-4c28-8437-4d54eb138bd7/volumes" Dec 01 09:28:37 crc kubenswrapper[4744]: I1201 09:28:37.011781 4744 scope.go:117] "RemoveContainer" containerID="e5935e70c30cab11e38e2f1f2ba798c26ac4416de72350dad59db62ac3aa51ae" Dec 01 09:28:37 crc kubenswrapper[4744]: I1201 09:28:37.011922 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/crc-debug-vzq7b" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.397274 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/init-config-reloader/0.log" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.557877 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/alertmanager/0.log" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.572310 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/init-config-reloader/0.log" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.635992 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_0c022f75-f489-4eee-a701-f19067fe78a2/config-reloader/0.log" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.812858 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-api/0.log" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.829822 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-listener/0.log" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.838180 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-evaluator/0.log" Dec 01 09:29:40 crc kubenswrapper[4744]: I1201 09:29:40.911963 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_00e7d3a7-76c8-4657-9168-5086d95c429a/aodh-notifier/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.001592 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b74d8fd6-9hczb_c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d/barbican-api/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.020463 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b74d8fd6-9hczb_c92bfcf4-c9e1-412b-a0e0-2aabeac15d5d/barbican-api-log/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.197620 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f8b9684d-5kjvj_02654eb7-3943-40d1-9145-04bcd38b8344/barbican-keystone-listener/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.238782 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f8b9684d-5kjvj_02654eb7-3943-40d1-9145-04bcd38b8344/barbican-keystone-listener-log/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.367506 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b8696b95c-fb5g6_a23750f1-c129-46b5-9c30-c6fa8ee54320/barbican-worker/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.374770 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b8696b95c-fb5g6_a23750f1-c129-46b5-9c30-c6fa8ee54320/barbican-worker-log/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.507334 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-rkg7c_0546ddaa-13e4-481a-b1f1-a2fe6aa505f3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.617661 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/ceilometer-central-agent/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.705505 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/proxy-httpd/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.715332 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/ceilometer-notification-agent/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.767552 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_999f566b-3f04-4cfd-9ff8-be798f51635a/sg-core/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.925019 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_08db63f5-2255-45a5-b2c2-b5baaed9fbca/cinder-api-log/0.log" Dec 01 09:29:41 crc kubenswrapper[4744]: I1201 09:29:41.969549 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_08db63f5-2255-45a5-b2c2-b5baaed9fbca/cinder-api/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.113894 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23516ec1-d13c-4dfd-b20f-207ba2bb0d26/cinder-scheduler/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.148823 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23516ec1-d13c-4dfd-b20f-207ba2bb0d26/probe/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.241017 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k4766_1f3aaa29-4bf7-4f79-aae9-7e90d131703d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.360649 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-j5b7f_e325084e-1abd-4b89-99f7-ec9e4f6f55ef/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.468794 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-5nlp9_dee5122c-f685-44a1-acd0-9fcfe3968bb7/init/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.642658 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-5nlp9_dee5122c-f685-44a1-acd0-9fcfe3968bb7/init/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.668052 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-5nlp9_dee5122c-f685-44a1-acd0-9fcfe3968bb7/dnsmasq-dns/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.676271 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ghbxl_2b675722-4753-4de2-8bd8-aa37599eb53e/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.848502 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a52e8d56-63ad-466d-a7e9-57c75c346a1c/glance-httpd/0.log" Dec 01 09:29:42 crc kubenswrapper[4744]: I1201 09:29:42.893531 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a52e8d56-63ad-466d-a7e9-57c75c346a1c/glance-log/0.log" Dec 01 09:29:43 crc kubenswrapper[4744]: I1201 09:29:43.038854 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7b5c1409-3b79-4680-beb3-14ffcf0fb055/glance-httpd/0.log" Dec 01 09:29:43 crc kubenswrapper[4744]: I1201 09:29:43.085892 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7b5c1409-3b79-4680-beb3-14ffcf0fb055/glance-log/0.log" Dec 01 09:29:43 crc kubenswrapper[4744]: I1201 09:29:43.577972 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5bf9f8b584-7k2r7_11429dc7-0531-4503-b66e-f6ecef318ec0/heat-engine/0.log" Dec 01 09:29:43 crc kubenswrapper[4744]: I1201 09:29:43.584897 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5c9d7d48-5qkmn_a60bb987-1492-435c-9f53-17c3820ce1a0/heat-api/0.log" Dec 01 09:29:43 crc kubenswrapper[4744]: I1201 09:29:43.679195 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68d7bb6659-xxvcz_6abf6646-56a2-4700-9a37-9474036b8865/heat-cfnapi/0.log" Dec 01 09:29:43 crc kubenswrapper[4744]: I1201 09:29:43.738065 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lhpn7_cbfdfe61-094b-48f5-a450-0fcac89b8e8a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:43 crc kubenswrapper[4744]: I1201 09:29:43.922336 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fttwk_b2c369ea-0c26-424a-af22-996d64ecf25d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:44 crc kubenswrapper[4744]: I1201 09:29:44.093254 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-77497cdd8f-s4w2q_d2a839b7-c745-45e3-9984-3f403dbf6fce/keystone-api/0.log" Dec 01 09:29:44 crc kubenswrapper[4744]: I1201 09:29:44.174430 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_13bbf1b1-4680-4390-99ed-e0a54dd6d6fe/kube-state-metrics/0.log" Dec 01 09:29:44 crc kubenswrapper[4744]: I1201 09:29:44.273854 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29409661-lht94_ee0b29aa-22fe-412e-ad54-43db39412cec/keystone-cron/0.log" Dec 01 09:29:44 crc kubenswrapper[4744]: I1201 09:29:44.329395 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pwsxm_b9ea1f1e-e881-4097-bd64-59ad7d48d552/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:44 crc kubenswrapper[4744]: I1201 09:29:44.673920 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-685cbf69b9-jrc9r_09924952-c000-422a-b471-e738246ae161/neutron-httpd/0.log" Dec 01 09:29:44 crc kubenswrapper[4744]: I1201 09:29:44.697434 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-685cbf69b9-jrc9r_09924952-c000-422a-b471-e738246ae161/neutron-api/0.log" Dec 01 09:29:44 crc kubenswrapper[4744]: I1201 09:29:44.903260 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-d5zjp_ede60eff-c329-466b-a25d-7a309d424bf3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:45 crc kubenswrapper[4744]: I1201 09:29:45.171734 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ec302fdc-d505-4f9c-b403-14643c501988/nova-api-log/0.log" Dec 01 09:29:45 crc kubenswrapper[4744]: I1201 09:29:45.391982 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d4ab1910-715b-4982-9903-453336b14b78/nova-cell0-conductor-conductor/0.log" Dec 01 09:29:45 crc kubenswrapper[4744]: I1201 09:29:45.583388 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0ae053c1-f41c-4bbb-bae8-66e1d3d891b7/nova-cell1-conductor-conductor/0.log" Dec 01 09:29:45 crc kubenswrapper[4744]: I1201 09:29:45.601533 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ec302fdc-d505-4f9c-b403-14643c501988/nova-api-api/0.log" Dec 01 09:29:45 crc kubenswrapper[4744]: I1201 09:29:45.735178 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c9f8476b-a1a9-42b5-ba40-95be5413d23b/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 09:29:45 crc kubenswrapper[4744]: I1201 09:29:45.888791 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8rpw6_da76f156-fbfc-46fe-b11d-a71a0bfd280d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:46 crc kubenswrapper[4744]: I1201 09:29:46.082528 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7bfdbce6-2016-4da0-8d18-300bbc3e92fb/nova-metadata-log/0.log" Dec 01 09:29:46 crc kubenswrapper[4744]: I1201 09:29:46.295758 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_97f2638c-75e4-4b25-be41-16e6b261c1b3/nova-scheduler-scheduler/0.log" Dec 01 09:29:46 crc kubenswrapper[4744]: I1201 09:29:46.366149 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c31db90c-376f-45b5-9d13-73bcf4962135/mysql-bootstrap/0.log" Dec 01 09:29:46 crc kubenswrapper[4744]: I1201 09:29:46.993908 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c31db90c-376f-45b5-9d13-73bcf4962135/mysql-bootstrap/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.007209 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c31db90c-376f-45b5-9d13-73bcf4962135/galera/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.225474 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0b711f28-53bd-46e8-9a86-44be43bd78d3/mysql-bootstrap/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.444728 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0b711f28-53bd-46e8-9a86-44be43bd78d3/mysql-bootstrap/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.462947 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0b711f28-53bd-46e8-9a86-44be43bd78d3/galera/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.565722 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7bfdbce6-2016-4da0-8d18-300bbc3e92fb/nova-metadata-metadata/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.644122 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_55446f52-3cf3-4ac2-9ad1-7f35c5508697/openstackclient/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.707137 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4nncr_efccc7e5-6075-43fb-b36e-15c4734e8f97/openstack-network-exporter/0.log" Dec 01 09:29:47 crc kubenswrapper[4744]: I1201 09:29:47.852574 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovsdb-server-init/0.log" Dec 01 09:29:48 crc kubenswrapper[4744]: I1201 09:29:48.050814 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovs-vswitchd/0.log" Dec 01 09:29:48 crc kubenswrapper[4744]: I1201 09:29:48.094250 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovsdb-server/0.log" Dec 01 09:29:48 crc kubenswrapper[4744]: I1201 09:29:48.095806 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fzxwj_2b14f48b-6456-4329-97fa-e81ec05e2644/ovsdb-server-init/0.log" Dec 01 09:29:48 crc kubenswrapper[4744]: I1201 09:29:48.358117 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pgp2f_ac7e9e72-1bca-42aa-a062-e9743651f2aa/ovn-controller/0.log" Dec 01 09:29:48 crc kubenswrapper[4744]: I1201 09:29:48.978594 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7v6p7_25afa440-2166-4a7f-a6b9-8d4dc2c7087c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.051137 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_551c38c7-428c-427a-b9b3-1322157a08df/openstack-network-exporter/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.218472 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_551c38c7-428c-427a-b9b3-1322157a08df/ovn-northd/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.296316 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ebb51f-ba73-417a-b6a8-ff31ff555108/openstack-network-exporter/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.355842 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6ebb51f-ba73-417a-b6a8-ff31ff555108/ovsdbserver-nb/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.549271 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_916bb441-60a0-4452-b2cf-5b98fb9aea9f/openstack-network-exporter/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.615119 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_916bb441-60a0-4452-b2cf-5b98fb9aea9f/ovsdbserver-sb/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.751348 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86964bb7b8-4bgsm_b7d2acd7-91fd-4cc6-a9de-f7a3fb905192/placement-api/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.850164 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-86964bb7b8-4bgsm_b7d2acd7-91fd-4cc6-a9de-f7a3fb905192/placement-log/0.log" Dec 01 09:29:49 crc kubenswrapper[4744]: I1201 09:29:49.870940 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/init-config-reloader/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.052895 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/init-config-reloader/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.069997 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/config-reloader/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.096818 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/prometheus/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.099679 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8243e221-4e33-4bd8-bc4b-67f0e0a0d170/thanos-sidecar/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.293064 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e4a93bac-aabe-4109-950f-14cb6ac634dd/setup-container/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.461471 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e4a93bac-aabe-4109-950f-14cb6ac634dd/setup-container/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.518296 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e4a93bac-aabe-4109-950f-14cb6ac634dd/rabbitmq/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.657076 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_93e40ea3-265b-44a9-ada5-5b583201148d/setup-container/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.892085 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-blhmz_de4a3818-81e1-4735-b58d-6d1bb9df5ee2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:50 crc kubenswrapper[4744]: I1201 09:29:50.953574 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_93e40ea3-265b-44a9-ada5-5b583201148d/setup-container/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.062722 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-mn867_e59297bb-e715-4a12-b214-70d9aa5c4757/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.317377 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p5fmz_75d5d4aa-8bf9-4042-b9e2-dfc02af5c1c6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.474192 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-25vvz_dbf61995-97c3-420f-a591-5867f9192102/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.533960 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_93e40ea3-265b-44a9-ada5-5b583201148d/rabbitmq/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.662710 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-g5snq_1e08d86e-f21d-42a3-8d93-be22013b97ac/ssh-known-hosts-edpm-deployment/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.862708 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696c89c747-77nz8_2e5f2594-2a7b-4c08-ac37-3e258f99fbb4/proxy-server/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.966735 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696c89c747-77nz8_2e5f2594-2a7b-4c08-ac37-3e258f99fbb4/proxy-httpd/0.log" Dec 01 09:29:51 crc kubenswrapper[4744]: I1201 09:29:51.984461 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-l9b2z_4f171eaa-706a-47ae-be94-1d1cbd636df1/swift-ring-rebalance/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.199765 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-auditor/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.228348 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-reaper/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.324703 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-replicator/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.375100 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/account-server/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.415618 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-auditor/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.452668 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-replicator/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.554343 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-server/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.608791 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/container-updater/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.639746 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-expirer/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.753000 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-auditor/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.793149 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-replicator/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.834688 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-updater/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.850171 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/object-server/0.log" Dec 01 09:29:52 crc kubenswrapper[4744]: I1201 09:29:52.976212 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/rsync/0.log" Dec 01 09:29:53 crc kubenswrapper[4744]: I1201 09:29:53.009754 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_424071bb-0b84-421b-9858-016c60e89b30/swift-recon-cron/0.log" Dec 01 09:29:53 crc kubenswrapper[4744]: I1201 09:29:53.142226 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s24rv_ffbaa140-8ee6-4639-9524-96460ffe5a49/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:29:53 crc kubenswrapper[4744]: I1201 09:29:53.229086 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9hsxt_25cf1554-6aff-4169-ae0e-a6b6979ad45e/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.186609 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p"] Dec 01 09:30:00 crc kubenswrapper[4744]: E1201 09:30:00.187897 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f96cd1-f706-4c28-8437-4d54eb138bd7" containerName="container-00" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.187919 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f96cd1-f706-4c28-8437-4d54eb138bd7" containerName="container-00" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.188331 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f96cd1-f706-4c28-8437-4d54eb138bd7" containerName="container-00" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.189728 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.192509 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.193728 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.197733 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p"] Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.356857 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn969\" (UniqueName: \"kubernetes.io/projected/99df622e-7749-4acd-aa05-139d3e4af9e6-kube-api-access-pn969\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.357163 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df622e-7749-4acd-aa05-139d3e4af9e6-secret-volume\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.357241 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df622e-7749-4acd-aa05-139d3e4af9e6-config-volume\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.458485 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df622e-7749-4acd-aa05-139d3e4af9e6-secret-volume\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.458549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df622e-7749-4acd-aa05-139d3e4af9e6-config-volume\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.458580 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn969\" (UniqueName: \"kubernetes.io/projected/99df622e-7749-4acd-aa05-139d3e4af9e6-kube-api-access-pn969\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.459819 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df622e-7749-4acd-aa05-139d3e4af9e6-config-volume\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.464994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df622e-7749-4acd-aa05-139d3e4af9e6-secret-volume\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.474777 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn969\" (UniqueName: \"kubernetes.io/projected/99df622e-7749-4acd-aa05-139d3e4af9e6-kube-api-access-pn969\") pod \"collect-profiles-29409690-mgl8p\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:00 crc kubenswrapper[4744]: I1201 09:30:00.520399 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:01 crc kubenswrapper[4744]: I1201 09:30:01.052991 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p"] Dec 01 09:30:01 crc kubenswrapper[4744]: I1201 09:30:01.873796 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" event={"ID":"99df622e-7749-4acd-aa05-139d3e4af9e6","Type":"ContainerStarted","Data":"2d9107386ea2f2f6f6cfe99f4526c50118b34b2ef9c067f4cae41b5e49f4156e"} Dec 01 09:30:01 crc kubenswrapper[4744]: I1201 09:30:01.874304 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" event={"ID":"99df622e-7749-4acd-aa05-139d3e4af9e6","Type":"ContainerStarted","Data":"9a7a43e3aa1ed113fc900866a84906dfbead1c568a0b51415fc2e3dab54b7bc8"} Dec 01 09:30:01 crc kubenswrapper[4744]: I1201 09:30:01.906335 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" podStartSLOduration=1.906313602 podStartE2EDuration="1.906313602s" podCreationTimestamp="2025-12-01 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:30:01.895211127 +0000 UTC m=+4333.884269048" watchObservedRunningTime="2025-12-01 09:30:01.906313602 +0000 UTC m=+4333.895371523" Dec 01 09:30:02 crc kubenswrapper[4744]: I1201 09:30:02.497252 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0153c296-2e70-43fa-90ae-bb007406d2c0/memcached/0.log" Dec 01 09:30:02 crc kubenswrapper[4744]: I1201 09:30:02.882290 4744 generic.go:334] "Generic (PLEG): container finished" podID="99df622e-7749-4acd-aa05-139d3e4af9e6" containerID="2d9107386ea2f2f6f6cfe99f4526c50118b34b2ef9c067f4cae41b5e49f4156e" exitCode=0 Dec 01 09:30:02 crc kubenswrapper[4744]: I1201 09:30:02.882331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" event={"ID":"99df622e-7749-4acd-aa05-139d3e4af9e6","Type":"ContainerDied","Data":"2d9107386ea2f2f6f6cfe99f4526c50118b34b2ef9c067f4cae41b5e49f4156e"} Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.287693 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.439092 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df622e-7749-4acd-aa05-139d3e4af9e6-secret-volume\") pod \"99df622e-7749-4acd-aa05-139d3e4af9e6\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.439147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df622e-7749-4acd-aa05-139d3e4af9e6-config-volume\") pod \"99df622e-7749-4acd-aa05-139d3e4af9e6\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.439354 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn969\" (UniqueName: \"kubernetes.io/projected/99df622e-7749-4acd-aa05-139d3e4af9e6-kube-api-access-pn969\") pod \"99df622e-7749-4acd-aa05-139d3e4af9e6\" (UID: \"99df622e-7749-4acd-aa05-139d3e4af9e6\") " Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.440089 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99df622e-7749-4acd-aa05-139d3e4af9e6-config-volume" (OuterVolumeSpecName: "config-volume") pod "99df622e-7749-4acd-aa05-139d3e4af9e6" (UID: "99df622e-7749-4acd-aa05-139d3e4af9e6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.445195 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99df622e-7749-4acd-aa05-139d3e4af9e6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "99df622e-7749-4acd-aa05-139d3e4af9e6" (UID: "99df622e-7749-4acd-aa05-139d3e4af9e6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.447669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99df622e-7749-4acd-aa05-139d3e4af9e6-kube-api-access-pn969" (OuterVolumeSpecName: "kube-api-access-pn969") pod "99df622e-7749-4acd-aa05-139d3e4af9e6" (UID: "99df622e-7749-4acd-aa05-139d3e4af9e6"). InnerVolumeSpecName "kube-api-access-pn969". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.541250 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn969\" (UniqueName: \"kubernetes.io/projected/99df622e-7749-4acd-aa05-139d3e4af9e6-kube-api-access-pn969\") on node \"crc\" DevicePath \"\"" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.541284 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99df622e-7749-4acd-aa05-139d3e4af9e6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.541293 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99df622e-7749-4acd-aa05-139d3e4af9e6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.899819 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" event={"ID":"99df622e-7749-4acd-aa05-139d3e4af9e6","Type":"ContainerDied","Data":"9a7a43e3aa1ed113fc900866a84906dfbead1c568a0b51415fc2e3dab54b7bc8"} Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.900235 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a7a43e3aa1ed113fc900866a84906dfbead1c568a0b51415fc2e3dab54b7bc8" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.899870 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409690-mgl8p" Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.963551 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5"] Dec 01 09:30:04 crc kubenswrapper[4744]: I1201 09:30:04.973568 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-g7lr5"] Dec 01 09:30:06 crc kubenswrapper[4744]: I1201 09:30:06.300729 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb169184-182d-416e-89e2-cafb9251f4d1" path="/var/lib/kubelet/pods/fb169184-182d-416e-89e2-cafb9251f4d1/volumes" Dec 01 09:30:18 crc kubenswrapper[4744]: I1201 09:30:18.634239 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:30:18 crc kubenswrapper[4744]: I1201 09:30:18.634787 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:30:21 crc kubenswrapper[4744]: I1201 09:30:21.762940 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/util/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.136282 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/pull/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.160434 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/util/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.188068 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/pull/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.387054 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/extract/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.387632 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/pull/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.407883 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9d8a48dc3f9662ac3564cf1fbbc1c83727c494fbc7abe3aa11036ffb68h5t9p_d667dc57-45ab-4792-ac76-f7de322dbb88/util/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.597738 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-wf77m_7bc8fb90-3b43-43f8-8df4-128c639ad76a/kube-rbac-proxy/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.618700 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-wf77m_7bc8fb90-3b43-43f8-8df4-128c639ad76a/manager/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.639476 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-lkltp_d9f50601-c7cd-4db6-82dd-9581a5dc8a10/kube-rbac-proxy/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.853002 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-8zfjd_2fbe6884-845e-4c7d-aff5-0deb483f4d19/manager/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.873815 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-lkltp_d9f50601-c7cd-4db6-82dd-9581a5dc8a10/manager/0.log" Dec 01 09:30:22 crc kubenswrapper[4744]: I1201 09:30:22.886209 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-8zfjd_2fbe6884-845e-4c7d-aff5-0deb483f4d19/kube-rbac-proxy/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.081836 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-kvcrb_f5fa46a8-32d1-4eda-89ec-04bdfd35ea78/kube-rbac-proxy/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.137921 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-kvcrb_f5fa46a8-32d1-4eda-89ec-04bdfd35ea78/manager/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.278455 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-7z2ml_badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a/kube-rbac-proxy/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.365071 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pv6f6_5f4c0ab8-4dee-4c39-b630-061580e95592/kube-rbac-proxy/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.369682 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-7z2ml_badf7ad5-ce9d-4fd7-9b86-ee3b4d535a6a/manager/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.481143 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pv6f6_5f4c0ab8-4dee-4c39-b630-061580e95592/manager/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.569969 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-dgxbb_ceb61e3f-5534-46ad-ba39-1fe0b07530f7/kube-rbac-proxy/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.777495 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-dgxbb_ceb61e3f-5534-46ad-ba39-1fe0b07530f7/manager/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.806041 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-gg7s7_214d7853-ed12-4624-82cd-6b279a827a13/manager/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.810873 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-gg7s7_214d7853-ed12-4624-82cd-6b279a827a13/kube-rbac-proxy/0.log" Dec 01 09:30:23 crc kubenswrapper[4744]: I1201 09:30:23.925305 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-7fn8v_8b89e5df-c38d-4465-aa74-b7c67e268a0d/kube-rbac-proxy/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.129141 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-7fn8v_8b89e5df-c38d-4465-aa74-b7c67e268a0d/manager/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.146067 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-tlszh_4cf33df0-316f-4acf-b105-8dc967fae633/kube-rbac-proxy/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.302614 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-tlszh_4cf33df0-316f-4acf-b105-8dc967fae633/manager/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.344526 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-wqvcv_f68be396-2f8f-4c90-99d3-bbe12b7d6c8e/manager/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.352230 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-wqvcv_f68be396-2f8f-4c90-99d3-bbe12b7d6c8e/kube-rbac-proxy/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.535359 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rbt96_b42166df-9137-4ddb-b9b0-33515fcab298/kube-rbac-proxy/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.545654 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rbt96_b42166df-9137-4ddb-b9b0-33515fcab298/manager/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.705367 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-q9dxh_5e8a8cd5-653f-4ef0-97c6-36b1a045f59a/kube-rbac-proxy/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.791134 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-q9dxh_5e8a8cd5-653f-4ef0-97c6-36b1a045f59a/manager/0.log" Dec 01 09:30:24 crc kubenswrapper[4744]: I1201 09:30:24.829065 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-htgsw_c0f26427-9c16-40b5-ad62-5a7161cba83c/kube-rbac-proxy/0.log" Dec 01 09:30:25 crc kubenswrapper[4744]: I1201 09:30:25.563726 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4khmps_83c42810-1f8a-4616-9544-db48c6020c85/kube-rbac-proxy/0.log" Dec 01 09:30:25 crc kubenswrapper[4744]: I1201 09:30:25.578151 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4khmps_83c42810-1f8a-4616-9544-db48c6020c85/manager/0.log" Dec 01 09:30:25 crc kubenswrapper[4744]: I1201 09:30:25.615211 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-htgsw_c0f26427-9c16-40b5-ad62-5a7161cba83c/manager/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.044635 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d866d68c9-2fxjt_3df0fc73-2812-42d2-9c7f-e1f33591e8a1/operator/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.046351 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-q498w_34cb91f3-fee7-4916-be15-b257cab23850/registry-server/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.154723 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-fjx5c_102b5fb4-119f-41ba-845e-1b691b08eb57/kube-rbac-proxy/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.309799 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-848rm_cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5/kube-rbac-proxy/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.321181 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-fjx5c_102b5fb4-119f-41ba-845e-1b691b08eb57/manager/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.400989 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-848rm_cfb2e5ee-ba79-4b5c-b7ab-537626ba09e5/manager/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.510615 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-ksvz2_c0a2f26e-7dc2-444f-aa8a-75e122512c98/operator/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.680911 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-ntg6w_b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad/kube-rbac-proxy/0.log" Dec 01 09:30:26 crc kubenswrapper[4744]: I1201 09:30:26.748457 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-ntg6w_b19a4ac2-1ca5-4af5-afe8-8d0122ca08ad/manager/0.log" Dec 01 09:30:27 crc kubenswrapper[4744]: I1201 09:30:27.460780 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/kube-rbac-proxy/0.log" Dec 01 09:30:27 crc kubenswrapper[4744]: I1201 09:30:27.460823 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-76f5b5884f-78kk6_717a3535-87a4-4d0a-b76d-a540a0cefa90/manager/0.log" Dec 01 09:30:27 crc kubenswrapper[4744]: I1201 09:30:27.562754 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-bxnxg_d602d407-4c7d-4057-a87b-b7a6fb426cfe/kube-rbac-proxy/0.log" Dec 01 09:30:27 crc kubenswrapper[4744]: I1201 09:30:27.688495 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-bxnxg_d602d407-4c7d-4057-a87b-b7a6fb426cfe/manager/0.log" Dec 01 09:30:27 crc kubenswrapper[4744]: I1201 09:30:27.694955 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-79bd889cc6-cprkk_7540e715-d2fd-4d78-a64c-e970f3da3c9b/manager/0.log" Dec 01 09:30:27 crc kubenswrapper[4744]: I1201 09:30:27.858890 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qrbr4_3eb6d34f-5dfb-4eb6-8d81-a7f11406531f/kube-rbac-proxy/0.log" Dec 01 09:30:27 crc kubenswrapper[4744]: I1201 09:30:27.861583 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qrbr4_3eb6d34f-5dfb-4eb6-8d81-a7f11406531f/manager/0.log" Dec 01 09:30:47 crc kubenswrapper[4744]: I1201 09:30:47.727234 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dmlw7_075bdaa2-f2c7-4c5a-b0ac-09da9c149141/control-plane-machine-set-operator/0.log" Dec 01 09:30:47 crc kubenswrapper[4744]: I1201 09:30:47.966469 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-l9mth_6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc/kube-rbac-proxy/0.log" Dec 01 09:30:47 crc kubenswrapper[4744]: I1201 09:30:47.979289 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-l9mth_6c6ffc7f-409c-4a41-b53d-38a6c85b5fcc/machine-api-operator/0.log" Dec 01 09:30:48 crc kubenswrapper[4744]: I1201 09:30:48.634026 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:30:48 crc kubenswrapper[4744]: I1201 09:30:48.634489 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:31:00 crc kubenswrapper[4744]: I1201 09:31:00.390051 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bjbkb_e05026bf-1db3-4a62-baff-a20dd3d8fd65/cert-manager-controller/0.log" Dec 01 09:31:00 crc kubenswrapper[4744]: I1201 09:31:00.507644 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-bp2pq_5bc4fc9c-b3ed-40f1-b951-8587803ba399/cert-manager-cainjector/0.log" Dec 01 09:31:00 crc kubenswrapper[4744]: I1201 09:31:00.583126 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-w864b_db8c6300-bdd5-4fe7-af23-ed9feb32b844/cert-manager-webhook/0.log" Dec 01 09:31:05 crc kubenswrapper[4744]: I1201 09:31:05.097603 4744 scope.go:117] "RemoveContainer" containerID="2d79de10d481a3d655b36f29f38936c7f27a4203fc8f6f987f1f81910e597c8e" Dec 01 09:31:12 crc kubenswrapper[4744]: I1201 09:31:12.906719 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-vrj4v_9d70fe1c-9023-42af-9b1e-c078b39399f6/nmstate-console-plugin/0.log" Dec 01 09:31:13 crc kubenswrapper[4744]: I1201 09:31:13.074112 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7r6x7_ea5a52f6-b418-4027-b6c0-3d57394d5fa5/nmstate-handler/0.log" Dec 01 09:31:13 crc kubenswrapper[4744]: I1201 09:31:13.163970 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lm574_5cdd7453-8251-4776-82bf-b931ba787d45/kube-rbac-proxy/0.log" Dec 01 09:31:13 crc kubenswrapper[4744]: I1201 09:31:13.205639 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lm574_5cdd7453-8251-4776-82bf-b931ba787d45/nmstate-metrics/0.log" Dec 01 09:31:13 crc kubenswrapper[4744]: I1201 09:31:13.281037 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-mmbdq_ce67e6e1-9203-47bd-abb6-1dfe9b552bca/nmstate-operator/0.log" Dec 01 09:31:13 crc kubenswrapper[4744]: I1201 09:31:13.419736 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-gq68c_5c7e45c3-5458-46ec-8c35-8e9776f6550e/nmstate-webhook/0.log" Dec 01 09:31:18 crc kubenswrapper[4744]: I1201 09:31:18.633972 4744 patch_prober.go:28] interesting pod/machine-config-daemon-g75sq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:31:18 crc kubenswrapper[4744]: I1201 09:31:18.634609 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:31:18 crc kubenswrapper[4744]: I1201 09:31:18.634657 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" Dec 01 09:31:18 crc kubenswrapper[4744]: I1201 09:31:18.635364 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353"} pod="openshift-machine-config-operator/machine-config-daemon-g75sq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:31:18 crc kubenswrapper[4744]: I1201 09:31:18.635431 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerName="machine-config-daemon" containerID="cri-o://373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" gracePeriod=600 Dec 01 09:31:18 crc kubenswrapper[4744]: E1201 09:31:18.759134 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:31:19 crc kubenswrapper[4744]: I1201 09:31:19.631884 4744 generic.go:334] "Generic (PLEG): container finished" podID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" exitCode=0 Dec 01 09:31:19 crc kubenswrapper[4744]: I1201 09:31:19.632019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerDied","Data":"373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353"} Dec 01 09:31:19 crc kubenswrapper[4744]: I1201 09:31:19.632206 4744 scope.go:117] "RemoveContainer" containerID="5c80c3abc5567fadc0827552b4cdbd5e7a3c17172db5d808070fd49e780b1417" Dec 01 09:31:19 crc kubenswrapper[4744]: I1201 09:31:19.632880 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:31:19 crc kubenswrapper[4744]: E1201 09:31:19.633181 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.389134 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bwdwl_edc692c2-6d9e-42fd-b54a-f856dd9f2ed1/kube-rbac-proxy/0.log" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.497778 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bwdwl_edc692c2-6d9e-42fd-b54a-f856dd9f2ed1/controller/0.log" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.592204 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-c6jzq_481b9c20-e688-4def-a63b-57038302c23c/frr-k8s-webhook-server/0.log" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.666197 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.838329 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.847144 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.852969 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:31:28 crc kubenswrapper[4744]: I1201 09:31:28.867319 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.055148 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.060296 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.071558 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.098390 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.249850 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-frr-files/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.310689 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/controller/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.313736 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-metrics/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.315419 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/cp-reloader/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.485613 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/kube-rbac-proxy/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.504955 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/kube-rbac-proxy-frr/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.529518 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/frr-metrics/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.719642 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79b5ddfc8f-x7dgg_6d5bd225-d8b4-4836-bb19-7e5c0fd540ae/manager/0.log" Dec 01 09:31:29 crc kubenswrapper[4744]: I1201 09:31:29.723025 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/reloader/0.log" Dec 01 09:31:30 crc kubenswrapper[4744]: I1201 09:31:30.027612 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68f89dd57b-kgd5g_84d30a9f-3fb4-4a8c-bc3a-cf9a45720d1d/webhook-server/0.log" Dec 01 09:31:30 crc kubenswrapper[4744]: I1201 09:31:30.238225 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6vzqh_3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c/kube-rbac-proxy/0.log" Dec 01 09:31:30 crc kubenswrapper[4744]: I1201 09:31:30.744596 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6vzqh_3be29ff8-9ae8-41be-bd0b-9a0ba9ce747c/speaker/0.log" Dec 01 09:31:31 crc kubenswrapper[4744]: I1201 09:31:31.122295 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wqpdx_33198d7f-46c9-41db-a7ce-4916d55561f2/frr/0.log" Dec 01 09:31:34 crc kubenswrapper[4744]: I1201 09:31:34.285165 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:31:34 crc kubenswrapper[4744]: E1201 09:31:34.286013 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.146246 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/util/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.336120 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/pull/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.403880 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/pull/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.408614 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/util/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.573488 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/pull/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.588878 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/util/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.589565 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f755q2_34673b73-0a10-419d-8604-d8f595325c05/extract/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.763878 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/util/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.907125 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/util/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.928907 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/pull/0.log" Dec 01 09:31:43 crc kubenswrapper[4744]: I1201 09:31:43.934085 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/pull/0.log" Dec 01 09:31:44 crc kubenswrapper[4744]: I1201 09:31:44.113276 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/util/0.log" Dec 01 09:31:44 crc kubenswrapper[4744]: I1201 09:31:44.114039 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/pull/0.log" Dec 01 09:31:44 crc kubenswrapper[4744]: I1201 09:31:44.125945 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210c5nvx_e4c3c3a6-d3d1-4461-9447-455fa3b5b3bb/extract/0.log" Dec 01 09:31:44 crc kubenswrapper[4744]: I1201 09:31:44.269620 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/util/0.log" Dec 01 09:31:44 crc kubenswrapper[4744]: I1201 09:31:44.456573 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/util/0.log" Dec 01 09:31:44 crc kubenswrapper[4744]: I1201 09:31:44.461699 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/pull/0.log" Dec 01 09:31:44 crc kubenswrapper[4744]: I1201 09:31:44.493111 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/pull/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.050192 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/util/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.085125 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/pull/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.135202 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838zhsb_d2dcf1e0-8bdb-41da-a406-0f9041c4c6c4/extract/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.226501 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-utilities/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.373422 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-content/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.392725 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-content/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.399812 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-utilities/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.591300 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-utilities/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.594268 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/extract-content/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.820102 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-utilities/0.log" Dec 01 09:31:45 crc kubenswrapper[4744]: I1201 09:31:45.993684 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-utilities/0.log" Dec 01 09:31:46 crc kubenswrapper[4744]: I1201 09:31:46.024880 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-content/0.log" Dec 01 09:31:46 crc kubenswrapper[4744]: I1201 09:31:46.077602 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-content/0.log" Dec 01 09:31:46 crc kubenswrapper[4744]: I1201 09:31:46.243627 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m5kzr_bbeb5e92-32fb-4db8-80bf-e7617fdd4e67/registry-server/0.log" Dec 01 09:31:46 crc kubenswrapper[4744]: I1201 09:31:46.266201 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-content/0.log" Dec 01 09:31:46 crc kubenswrapper[4744]: I1201 09:31:46.327740 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/extract-utilities/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.010396 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x49gs_31f8c61a-d957-4ad9-bdf9-16670c20020e/registry-server/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.112754 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-99jdl_94e841b2-c9ed-43c3-b96c-07661415a958/marketplace-operator/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.123175 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-utilities/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.243168 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-content/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.247609 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-utilities/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.301007 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-content/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.425819 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-content/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.470374 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/extract-utilities/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.476924 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-utilities/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.586158 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wd2vs_a5519f09-e6ba-4df8-8d28-d014f65fff6f/registry-server/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.696277 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-utilities/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.702116 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-content/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.705155 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-content/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.832273 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-utilities/0.log" Dec 01 09:31:47 crc kubenswrapper[4744]: I1201 09:31:47.853383 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/extract-content/0.log" Dec 01 09:31:48 crc kubenswrapper[4744]: I1201 09:31:48.283846 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5lhd_40ca9d77-9a9a-41dc-92eb-46706b9b3456/registry-server/0.log" Dec 01 09:31:48 crc kubenswrapper[4744]: I1201 09:31:48.310815 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:31:48 crc kubenswrapper[4744]: E1201 09:31:48.311681 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:31:59 crc kubenswrapper[4744]: I1201 09:31:59.273540 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-jt269_d7b03d83-6bcb-4e4d-beb9-d4932e4fab91/prometheus-operator/0.log" Dec 01 09:31:59 crc kubenswrapper[4744]: I1201 09:31:59.468329 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c7545f77c-hqgtj_50194db7-90e4-4112-83df-f729548b59ca/prometheus-operator-admission-webhook/0.log" Dec 01 09:31:59 crc kubenswrapper[4744]: I1201 09:31:59.468479 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c7545f77c-ml2lp_92c3c976-90f4-4d13-9d88-348767b50827/prometheus-operator-admission-webhook/0.log" Dec 01 09:31:59 crc kubenswrapper[4744]: I1201 09:31:59.686801 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-m2tkm_1466d56b-9044-4bd8-97f3-02c1feafa96f/operator/0.log" Dec 01 09:31:59 crc kubenswrapper[4744]: I1201 09:31:59.739206 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-tk4rw_d48e874b-8560-49fa-9981-136edeec81f7/perses-operator/0.log" Dec 01 09:32:00 crc kubenswrapper[4744]: I1201 09:32:00.288766 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:32:00 crc kubenswrapper[4744]: E1201 09:32:00.289023 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:32:13 crc kubenswrapper[4744]: I1201 09:32:13.285707 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:32:13 crc kubenswrapper[4744]: E1201 09:32:13.286441 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:32:27 crc kubenswrapper[4744]: I1201 09:32:27.285532 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:32:27 crc kubenswrapper[4744]: E1201 09:32:27.286131 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:32:42 crc kubenswrapper[4744]: I1201 09:32:42.288073 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:32:42 crc kubenswrapper[4744]: E1201 09:32:42.289003 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:32:57 crc kubenswrapper[4744]: I1201 09:32:57.285912 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:32:57 crc kubenswrapper[4744]: E1201 09:32:57.286533 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:33:11 crc kubenswrapper[4744]: I1201 09:33:11.285488 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:33:11 crc kubenswrapper[4744]: E1201 09:33:11.286108 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:33:26 crc kubenswrapper[4744]: I1201 09:33:26.285112 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:33:26 crc kubenswrapper[4744]: E1201 09:33:26.286072 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:33:32 crc kubenswrapper[4744]: I1201 09:33:32.998478 4744 generic.go:334] "Generic (PLEG): container finished" podID="4caf2e95-e84d-4818-bdea-651a449165fa" containerID="1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0" exitCode=0 Dec 01 09:33:32 crc kubenswrapper[4744]: I1201 09:33:32.998570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4cr8/must-gather-kg98g" event={"ID":"4caf2e95-e84d-4818-bdea-651a449165fa","Type":"ContainerDied","Data":"1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0"} Dec 01 09:33:33 crc kubenswrapper[4744]: I1201 09:33:32.999611 4744 scope.go:117] "RemoveContainer" containerID="1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0" Dec 01 09:33:33 crc kubenswrapper[4744]: I1201 09:33:33.215368 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x4cr8_must-gather-kg98g_4caf2e95-e84d-4818-bdea-651a449165fa/gather/0.log" Dec 01 09:33:38 crc kubenswrapper[4744]: I1201 09:33:38.295437 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:33:38 crc kubenswrapper[4744]: E1201 09:33:38.296366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:33:44 crc kubenswrapper[4744]: I1201 09:33:44.361041 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x4cr8/must-gather-kg98g"] Dec 01 09:33:44 crc kubenswrapper[4744]: I1201 09:33:44.361856 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x4cr8/must-gather-kg98g" podUID="4caf2e95-e84d-4818-bdea-651a449165fa" containerName="copy" containerID="cri-o://b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa" gracePeriod=2 Dec 01 09:33:44 crc kubenswrapper[4744]: I1201 09:33:44.374911 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x4cr8/must-gather-kg98g"] Dec 01 09:33:44 crc kubenswrapper[4744]: I1201 09:33:44.913421 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x4cr8_must-gather-kg98g_4caf2e95-e84d-4818-bdea-651a449165fa/copy/0.log" Dec 01 09:33:44 crc kubenswrapper[4744]: I1201 09:33:44.914344 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.115055 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4caf2e95-e84d-4818-bdea-651a449165fa-must-gather-output\") pod \"4caf2e95-e84d-4818-bdea-651a449165fa\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.115355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rbk\" (UniqueName: \"kubernetes.io/projected/4caf2e95-e84d-4818-bdea-651a449165fa-kube-api-access-w9rbk\") pod \"4caf2e95-e84d-4818-bdea-651a449165fa\" (UID: \"4caf2e95-e84d-4818-bdea-651a449165fa\") " Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.121117 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4caf2e95-e84d-4818-bdea-651a449165fa-kube-api-access-w9rbk" (OuterVolumeSpecName: "kube-api-access-w9rbk") pod "4caf2e95-e84d-4818-bdea-651a449165fa" (UID: "4caf2e95-e84d-4818-bdea-651a449165fa"). InnerVolumeSpecName "kube-api-access-w9rbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.158701 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x4cr8_must-gather-kg98g_4caf2e95-e84d-4818-bdea-651a449165fa/copy/0.log" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.159317 4744 generic.go:334] "Generic (PLEG): container finished" podID="4caf2e95-e84d-4818-bdea-651a449165fa" containerID="b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa" exitCode=143 Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.159374 4744 scope.go:117] "RemoveContainer" containerID="b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.159475 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4cr8/must-gather-kg98g" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.222529 4744 scope.go:117] "RemoveContainer" containerID="1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.223640 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rbk\" (UniqueName: \"kubernetes.io/projected/4caf2e95-e84d-4818-bdea-651a449165fa-kube-api-access-w9rbk\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.296437 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4caf2e95-e84d-4818-bdea-651a449165fa-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4caf2e95-e84d-4818-bdea-651a449165fa" (UID: "4caf2e95-e84d-4818-bdea-651a449165fa"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.299623 4744 scope.go:117] "RemoveContainer" containerID="b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa" Dec 01 09:33:45 crc kubenswrapper[4744]: E1201 09:33:45.300002 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa\": container with ID starting with b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa not found: ID does not exist" containerID="b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.300107 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa"} err="failed to get container status \"b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa\": rpc error: code = NotFound desc = could not find container \"b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa\": container with ID starting with b46c1a98c20b88dbbd45dda3ebb1997e7407cd17fbacebbbb2e2130cfd464aaa not found: ID does not exist" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.300188 4744 scope.go:117] "RemoveContainer" containerID="1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0" Dec 01 09:33:45 crc kubenswrapper[4744]: E1201 09:33:45.300539 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0\": container with ID starting with 1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0 not found: ID does not exist" containerID="1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.300570 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0"} err="failed to get container status \"1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0\": rpc error: code = NotFound desc = could not find container \"1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0\": container with ID starting with 1d2a17dce5619d1b028145f30e16f94d7686f23a6e0e7215fc53a5b5964baff0 not found: ID does not exist" Dec 01 09:33:45 crc kubenswrapper[4744]: I1201 09:33:45.325912 4744 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4caf2e95-e84d-4818-bdea-651a449165fa-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 09:33:46 crc kubenswrapper[4744]: I1201 09:33:46.296246 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4caf2e95-e84d-4818-bdea-651a449165fa" path="/var/lib/kubelet/pods/4caf2e95-e84d-4818-bdea-651a449165fa/volumes" Dec 01 09:33:49 crc kubenswrapper[4744]: I1201 09:33:49.287073 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:33:49 crc kubenswrapper[4744]: E1201 09:33:49.287863 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:34:04 crc kubenswrapper[4744]: I1201 09:34:04.284900 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:34:04 crc kubenswrapper[4744]: E1201 09:34:04.285749 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:34:05 crc kubenswrapper[4744]: I1201 09:34:05.211936 4744 scope.go:117] "RemoveContainer" containerID="c860f7ca32eb09ce5eb8de5136c1d5a3078db250676188550ded140901c5f320" Dec 01 09:34:05 crc kubenswrapper[4744]: I1201 09:34:05.239888 4744 scope.go:117] "RemoveContainer" containerID="fdacc725fbe1e18841e6dc676e8c6b52e887afe2b0c7f1bd148577d53435b62e" Dec 01 09:34:05 crc kubenswrapper[4744]: I1201 09:34:05.268751 4744 scope.go:117] "RemoveContainer" containerID="2e9decf918bf945e13f4a3b6c38b927dfd65e5f3c9332f5ce93bf0379812c9ba" Dec 01 09:34:16 crc kubenswrapper[4744]: I1201 09:34:16.285210 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:34:16 crc kubenswrapper[4744]: E1201 09:34:16.286088 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:34:28 crc kubenswrapper[4744]: I1201 09:34:28.291238 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:34:28 crc kubenswrapper[4744]: E1201 09:34:28.292316 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:34:39 crc kubenswrapper[4744]: I1201 09:34:39.285067 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:34:39 crc kubenswrapper[4744]: E1201 09:34:39.285992 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:34:51 crc kubenswrapper[4744]: I1201 09:34:51.285150 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:34:51 crc kubenswrapper[4744]: E1201 09:34:51.285994 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:35:03 crc kubenswrapper[4744]: I1201 09:35:03.284726 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:35:03 crc kubenswrapper[4744]: E1201 09:35:03.285521 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:35:14 crc kubenswrapper[4744]: I1201 09:35:14.285822 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:35:14 crc kubenswrapper[4744]: E1201 09:35:14.286606 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:35:25 crc kubenswrapper[4744]: I1201 09:35:25.284862 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:35:25 crc kubenswrapper[4744]: E1201 09:35:25.285611 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:35:40 crc kubenswrapper[4744]: I1201 09:35:40.285942 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:35:40 crc kubenswrapper[4744]: E1201 09:35:40.286673 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:35:51 crc kubenswrapper[4744]: I1201 09:35:51.285090 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:35:51 crc kubenswrapper[4744]: E1201 09:35:51.285787 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:36:06 crc kubenswrapper[4744]: I1201 09:36:06.287592 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:36:06 crc kubenswrapper[4744]: E1201 09:36:06.288565 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g75sq_openshift-machine-config-operator(c8ecf18a-4cfa-4266-87ae-1a08097cb5d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" podUID="c8ecf18a-4cfa-4266-87ae-1a08097cb5d1" Dec 01 09:36:20 crc kubenswrapper[4744]: I1201 09:36:20.285520 4744 scope.go:117] "RemoveContainer" containerID="373f11a1b6fbec562eb7636ccff798205829a5478cee6efeb5e48ee0e8572353" Dec 01 09:36:20 crc kubenswrapper[4744]: I1201 09:36:20.794864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g75sq" event={"ID":"c8ecf18a-4cfa-4266-87ae-1a08097cb5d1","Type":"ContainerStarted","Data":"cca7bddba28cb59be68fe51a7be6d05ecc4485777e1f9bcfafe29c377ca0d158"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113260651024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113260652017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113247154016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113247154015460 5ustar corecore